Learning Partially Observable Markov Decision Processes Using Abstract Actions
Hamed Janzadeh
openalex +1 more source
Intelligent power control using deep neural networks and regularized learning for shipboard microgrid. [PDF]
Deng W +5 more
europepmc +1 more source
Moving beyond diagnostic labels in psychiatry: outcome-linked treatment modelling. [PDF]
Lyndon S.
europepmc +1 more source
Transformer-Based Soft Actor-Critic for UAV Path Planning in Precision Agriculture IoT Networks. [PDF]
Ge G, Sun M, Xue Y, Pavlova S.
europepmc +1 more source
Resilience driven EV coordination in multiple microgrids using distributed deep reinforcement learning. [PDF]
Wu Y, Cai T, Li X.
europepmc +1 more source
Modeling treatment of ischemic heart disease with partially observable Markov decision processes.
Miloš Hauskrecht, Hamish Fraser
openalex +1 more source
Deep multi-objective reinforcement learning for utility-based infrastructural maintenance optimization. [PDF]
van Remmerden J +5 more
europepmc +1 more source
Related searches:
Partially Observable Markov Decision Processes
2011In many applications the decision maker has only partial information about the state process, i.e. part of the state cannot be observed. Examples can be found in engineering, economics, statistics, speech recognition and learning theory among others. An important financial application is given when the drift of a stock price process is unobservable and
Nicole Bäuerle, Ulrich Rieder
openaire +2 more sources
Partially Observable Markov Decision Processes
2020This chapter covers Partially Observable Markov Decision Processes (POMDPs), that extend MDPs for when the state is not completely observable. After a general introduction to POMDPs, their formal representation and properties are described. The representation of the value function as a set of linear equations (\(\alpha -vectors\)) is presented via a ...
openaire +1 more source
Partially observed Markov decision processes with binomial observations
Operations Research Letters, 2013Abstract We consider partially observed Markov decision processes with control limits. We analytically show how the finite-horizon control limits are non-monotonic in (a) the time remaining and (b) the probability of obtaining a conforming unit. We also prove that the infinite-horizon control limit can be calculated by solving a finite set of linear ...
Tal Ben-Zvi, Abraham Grosfeld-Nir
openaire +1 more source

