Results 41 to 50 of about 937,585 (275)
Flowsheet generation through hierarchical reinforcement learning and graph neural networks
Abstract Process synthesis experiences a disruptive transformation accelerated by artificial intelligence. We propose a reinforcement learning algorithm for chemical process design based on a state‐of‐the‐art actor‐critic logic. Our proposed algorithm represents chemical processes as graphs and uses graph convolutional neural networks to learn from ...
Laura Stops+3 more
wiley +1 more source
Forward and Backward Bellman Equations Improve the Efficiency of the EM Algorithm for DEC-POMDP
Decentralized partially observable Markov decision process (DEC-POMDP) models sequential decision making problems by a team of agents. Since the planning of DEC-POMDP can be interpreted as the maximum likelihood estimation for the latent variable model ...
Takehiro Tottori, Tetsuya J. Kobayashi
doaj +1 more source
Hidden Markov Model Estimation-Based Q-learning for Partially Observable Markov Decision Process [PDF]
The objective is to study an on-line Hidden Markov model (HMM) estimation-based Q-learning algorithm for partially observable Markov decision process (POMDP) on finite state and action sets.
Hyung-Jin Yoon+2 more
semanticscholar +1 more source
Abstract Premise Approximately 14% of all fern species have physiologically active chlorophyllous spores that are much more short‐lived than the more common and dormant achlorophyllous spores. Most chlorophyllous‐spored species (70%) are epiphytes and account for almost 37% of all epiphytic ferns.
Daniela Mellado‐Mansilla+6 more
wiley +1 more source
Cost‐effectiveness of second‐line therapies in adults with chronic immune thrombocytopenia
Abstract Major options for second‐line therapy in adults with chronic immune thrombocytopenia (ITP) include splenectomy, rituximab, and thrombopoietin receptor agonists (TRAs). The American Society of Hematology guidelines recommend rituximab over splenectomy, TRAs over rituximab, and splenectomy or TRAs while noting a lack of evidence on the cost ...
George Goshua+5 more
wiley +1 more source
On the complexity of partially observed Markov decision processes
AbstractIn the paper we consider the complexity of constructing optimal policies (strategies) for some type of partially observed Markov decision processes. This particular case of the classical problem deals with finite stationary processes, and can be represented as constructing optimal strategies to reach target vertices from a starting vertex in a ...
Michel de Rougemont+2 more
openaire +2 more sources
Reinforcement Learning-Based Detection for State Estimation Under False Data Injection
We consider the problem of network security under false data injection attacks over wireless sensor networks.To resist the attacks which can inject false data into communication channels according to a certain probability, we formulate the online attack ...
Weiliang Jiang+5 more
doaj +1 more source
Could a phase model help to improve translational animal research?
Transferability of findings from animal research to the human organism is controversially discussed. We propose a phase model which refers to the well‐known process of clinical research in order to improve decision‐making regarding the evidence which is gained in translational animal research.
Benjamin Mayer+2 more
wiley +1 more source
An intelligent land vehicle utilizes onboard sensors to acquire observed states at a disorderly intersection. However, partial observation of the environment occurs due to sensor noise. This causes decision failure easily.
Lingli Yu+3 more
doaj +1 more source
In this article, we study the ground moving target tracking problem for a fixed-wing unmanned aerial vehicle equipped with a radar. This problem is formulated in a partially observable Markov process framework, which contains the following two parts: in ...
Yunyun Zhao+3 more
doaj +1 more source