Results 261 to 270 of about 937,585 (275)
Some of the next articles are maybe not open access.
Partially observable Markov decision processes for artificial intelligence
1995In this paper, we bring techniques from operations research to bear on the problem of choosing optimal actions in partially observable stochastic domains. In many cases, we have developed new ways of viewing the problem that are, perhaps, more consistent with the AI perspective. We begin by introducing the theory of Markov decision processes (Mdps) and
Michael L. Littman+2 more
openaire +3 more sources
, 2011
This paper proposes a generalised partially observable Markov decision process (POMDP) combining decision analysis and dynamic programming. The model is applied to a maintenance optimisation problem, even though the scope of its utility is far beyond ...
R. Faddoul, W. Raphael, A. Chateauneuf
semanticscholar +1 more source
This paper proposes a generalised partially observable Markov decision process (POMDP) combining decision analysis and dynamic programming. The model is applied to a maintenance optimisation problem, even though the scope of its utility is far beyond ...
R. Faddoul, W. Raphael, A. Chateauneuf
semanticscholar +1 more source
Reward revision for partially observed Markov decision processes
1985 24th IEEE Conference on Decision and Control, 1985We present an algorithm for accelerating a successive approximation solution procedure for the infinite horizon, expected discounted total reward partially observed Markov decision process (POMDP). This algorithm proceeds by simultaneously constructing and solving a sequence of dynamic programs, each of which is more computationally attractive than the
Chelsea C. White, William T. Scherer
openaire +2 more sources
International journal of automotive technology, 2021
Duy Quang Tran, Sang-Hoon Bae
semanticscholar +1 more source
Duy Quang Tran, Sang-Hoon Bae
semanticscholar +1 more source
Partially observed Markov decision processes (POMDPs)
2016A POMDP is a controlled HMM. Recall from ยง2.4 that an HMM consists of an X -state Markov chain { x k } observed via a noisy observation process { y k }. Figure 7.1 displays the schematic setup of a POMDP where the action u k affects the state and/or observation (sensing) process of the HMM.
openaire +2 more sources
Solution Procedures for Partially Observed Markov Decision Processes
Operations Research, 1989We present three algorithms to solve the infinite horizon, expected discounted total reward partially observed Markov decision process (POMDP). Each algorithm integrates a successive approximations algorithm for the POMDP due to A. Smallwood and E.
Chelsea C. White, William T. Scherer
openaire +2 more sources
Reliability Engineering & System Safety, 2022
Chunhui Guo, Zhenglin Liang
semanticscholar +1 more source
Chunhui Guo, Zhenglin Liang
semanticscholar +1 more source
Structural Results for Partially Observable Markov Decision Processes
Operations Research, 1979This paper examines monotonicity results for a fairly general class of partially observable Markov decision processes. When there are only two actual states in the system and when the actions taken are primarily intended to improve the system, rather than to inspect it, we give reasonable conditions which ensure that the optimal reward function and ...
openaire +1 more source