Results 281 to 290 of about 35,395 (320)
Some of the next articles are maybe not open access.

Approximations for Partially Observed Markov Decision Processes

2018
This chapter studies the finite-model approximation of discrete-time partially observed Markov decision process. We will find that by performing the standard reduction method, where one transforms a partially observed model to a belief-based fully observed model, we can apply and properly generalize the results in the preceding chapters to obtain ...
Saldı, Naci, Linder, T., Yüksel, S.
openaire   +2 more sources

Partially Observable Markov Decision Processes and Robotics

Annual Review of Control, Robotics, and Autonomous Systems, 2022
Planning under uncertainty is critical to robotics. The partially observable Markov decision process (POMDP) is a mathematical framework for such planning problems. POMDPs are powerful because of their careful quantification of the nondeterministic effects of actions and the partial observability of the states.
openaire   +1 more source

Competitive Markov decision processes with partial observation

2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583), 2005
We study a class of Markov decision processes (MDPs) in the infinite time horizon where the number of controllers is two and the observation information is allowed to be imperfect. Suppose the system, space and action space are both finite, and the controllers, having conflicting interests with each other, make decisions independently to seek their own
null Shun-Pin Hsu, A. Arapostathis
openaire   +1 more source

Structural Results for Partially Observable Markov Decision Processes

Operations Research, 1979
This paper examines monotonicity results for a fairly general class of partially observable Markov decision processes. When there are only two actual states in the system and when the actions taken are primarily intended to improve the system, rather than to inspect it, we give reasonable conditions which ensure that the optimal reward function and ...
openaire   +1 more source

Partially observable Markov decision processes for artificial intelligence

1995
In this paper, we bring techniques from operations research to bear on the problem of choosing optimal actions in partially observable stochastic domains. In many cases, we have developed new ways of viewing the problem that are, perhaps, more consistent with the AI perspective. We begin by introducing the theory of Markov decision processes (Mdps) and
Leslie Pack Kaelbling   +2 more
openaire   +1 more source

Partially Observed Markov Decision Processes

2016
Covering formulation, algorithms, and structural results, and linking theory to real-world applications in controlled sensing (including social learning, adaptive radars and sequential detection), this book focuses on the conceptual foundations of partially observed Markov decision processes (POMDPs).
openaire   +1 more source

Partially Observed Markov Decision Processes

Covering formulation, algorithms and structural results and linking theory to real-world applications in controlled sensing (including social learning, adaptive radars and sequential detection), this book focuses on the conceptual foundations of partially observed Markov decision processes (POMDPs).
openaire   +2 more sources

A Partially Observable Markov Decision Process with Lagged Information

Journal of the Operational Research Society, 1987
The paper considers a Markov decision model where instead of the real state S(t) at stage t one observes some state \(S_ c(t)\) at time t and some state \(S_{\ell}(t+1)\) at time \(t+1\) (simultaneously with \(S_ c(t+1))\). The authors tacitly assume some conditional independence between \(S_ c(t)\) and \(S_{\ell}(t+1)\) given S(t), i.e., (1) and (2 ...
KIM, SH Kim, Soung Hie   +1 more
openaire   +1 more source

Quasi-Deterministic Partially Observable Markov Decision Processes

2009
We study a subclass of pomdps, called quasi-deterministic pomdps (qDet- pomdps), characterized by deterministic actions and stochastic observations. While this framework does not model the same general problems as pomdps, they still capture a number of interesting and challenging problems and, in some cases, have interesting properties. By studying the
Camille Besse, Brahim Chaib-draa
openaire   +1 more source

Partially Observable Markov Decision Processes and Performance Sensitivity Analysis

IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 2008
The sensitivity-based optimization of Markov systems has become an increasingly important area. From the perspective of performance sensitivity analysis, policy-iteration algorithms and gradient estimation methods can be directly obtained for Markov decision processes (MDPs).
Yanjie, Li, Baoqun, Yin, Hongsheng, Xi
openaire   +2 more sources

Home - About - Disclaimer - Privacy