Results 31 to 40 of about 30,788 (321)

Partially observable Markov decision processes with partially observable random discount factors

open access: yesKybernetika, 2023
A discrete time Markov decision process with Borel state and action process and with discounted, unbounded one-step costs is considered. The value of the discount factor is Markov random and independent of the state process. The transition functions for both processes are explicitly given.
Martinez-Garcia, E. Everardo   +2 more
openaire   +2 more sources

Cost-Bounded Active Classification Using Partially Observable Markov Decision Processes [PDF]

open access: yes, 2018
Active classification, i.e., the sequential decision-making process aimed at data acquisition for classification purposes, arises naturally in many applications, including medical diagnosis, intrusion detection, and object tracking.
Ahmadi, Mohamadreza   +3 more
core   +2 more sources

On Anderson Acceleration for Partially Observable Markov Decision Processes [PDF]

open access: yes2021 60th IEEE Conference on Decision and Control (CDC), 2021
This paper proposes an accelerated method for approximately solving partially observable Markov decision process (POMDP) problems offline. Our method carefully combines two existing tools: Anderson acceleration (AA) and the fast informed bound (FIB) method.
Ermis, Melike   +2 more
openaire   +2 more sources

Partially observable Markov decision processes with partially observable random discount factors [PDF]

open access: yes, 2022
summary:This paper deals with a class of partially observable discounted Markov decision processes defined on Borel state and action spaces, under unbounded one-stage cost.
Martinez-Garcia, E. Everardo   +2 more
core   +1 more source

Wireless Body Area Network Control Policies for Energy-Efficient Health Monitoring

open access: yesSensors, 2021
Wireless body area networks (WBANs) have strong potential in the field of health monitoring. However, the energy consumption required for accurate monitoring determines the time between battery charges of the wearable sensors, which is a key performance ...
Yair Bar David   +5 more
doaj   +1 more source

Partially Observable Markov Decision Processes [PDF]

open access: yes, 2012
For reinforcement learning in environments in which an agent has access to a reliable state signal, methods based on the Markov decision process (MDP) have had many successes. In many problem domains, however, an agent suffers from limited sensing capabilities that preclude it from recovering a Markovian state signal from its perceptions. Extending the
openaire   +2 more sources

Blackwell optimality in Markov decision processes with partial observation [PDF]

open access: yesThe Annals of Statistics, 2002
zbMATH Open Web Interface contents unavailable due to conflicting licenses.
Rosenberg, Dinah   +2 more
openaire   +6 more sources

Decision Making with STPA through Markov Decision Process, a Theoretic Framework for Safe Human-Robot Collaboration

open access: yesApplied Sciences, 2021
During the last decades, collaborative robots capable of operating out of their cages are widely used in industry to assist humans in mundane and harsh manufacturing tasks. Although such robots are inherently safe by design, they are commonly accompanied
Angeliki Zacharaki   +2 more
doaj   +1 more source

Model-Free Deep Recurrent Q-Network Reinforcement Learning for Quantum Circuit Architectures Design

open access: yesQuantum Reports, 2022
Artificial intelligence (AI) technology leads to new insights into the manipulation of quantum systems in the Noisy Intermediate-Scale Quantum (NISQ) era.
Tomah Sogabe   +6 more
doaj   +1 more source

Multiple-Environment Markov Decision Processes [PDF]

open access: yes, 2014
We introduce Multi-Environment Markov Decision Processes (MEMDPs) which are MDPs with a set of probabilistic transition functions. The goal in a MEMDP is to synthesize a single controller with guaranteed performances against all environments even though ...
Raskin, Jean-François, Sankur, Ocan
core   +2 more sources

Home - About - Disclaimer - Privacy