Results 31 to 40 of about 30,788 (321)
Partially observable Markov decision processes with partially observable random discount factors
A discrete time Markov decision process with Borel state and action process and with discounted, unbounded one-step costs is considered. The value of the discount factor is Markov random and independent of the state process. The transition functions for both processes are explicitly given.
Martinez-Garcia, E. Everardo +2 more
openaire +2 more sources
Cost-Bounded Active Classification Using Partially Observable Markov Decision Processes [PDF]
Active classification, i.e., the sequential decision-making process aimed at data acquisition for classification purposes, arises naturally in many applications, including medical diagnosis, intrusion detection, and object tracking.
Ahmadi, Mohamadreza +3 more
core +2 more sources
On Anderson Acceleration for Partially Observable Markov Decision Processes [PDF]
This paper proposes an accelerated method for approximately solving partially observable Markov decision process (POMDP) problems offline. Our method carefully combines two existing tools: Anderson acceleration (AA) and the fast informed bound (FIB) method.
Ermis, Melike +2 more
openaire +2 more sources
Partially observable Markov decision processes with partially observable random discount factors [PDF]
summary:This paper deals with a class of partially observable discounted Markov decision processes defined on Borel state and action spaces, under unbounded one-stage cost.
Martinez-Garcia, E. Everardo +2 more
core +1 more source
Wireless Body Area Network Control Policies for Energy-Efficient Health Monitoring
Wireless body area networks (WBANs) have strong potential in the field of health monitoring. However, the energy consumption required for accurate monitoring determines the time between battery charges of the wearable sensors, which is a key performance ...
Yair Bar David +5 more
doaj +1 more source
Partially Observable Markov Decision Processes [PDF]
For reinforcement learning in environments in which an agent has access to a reliable state signal, methods based on the Markov decision process (MDP) have had many successes. In many problem domains, however, an agent suffers from limited sensing capabilities that preclude it from recovering a Markovian state signal from its perceptions. Extending the
openaire +2 more sources
Blackwell optimality in Markov decision processes with partial observation [PDF]
zbMATH Open Web Interface contents unavailable due to conflicting licenses.
Rosenberg, Dinah +2 more
openaire +6 more sources
During the last decades, collaborative robots capable of operating out of their cages are widely used in industry to assist humans in mundane and harsh manufacturing tasks. Although such robots are inherently safe by design, they are commonly accompanied
Angeliki Zacharaki +2 more
doaj +1 more source
Model-Free Deep Recurrent Q-Network Reinforcement Learning for Quantum Circuit Architectures Design
Artificial intelligence (AI) technology leads to new insights into the manipulation of quantum systems in the Noisy Intermediate-Scale Quantum (NISQ) era.
Tomah Sogabe +6 more
doaj +1 more source
Multiple-Environment Markov Decision Processes [PDF]
We introduce Multi-Environment Markov Decision Processes (MEMDPs) which are MDPs with a set of probabilistic transition functions. The goal in a MEMDP is to synthesize a single controller with guaranteed performances against all environments even though ...
Raskin, Jean-François, Sankur, Ocan
core +2 more sources

