Making virtual learning environment more intelligent: application of Markov decision process
Suppose there exist a Virtual Learning Environment in which agent plays a role of the teacher. With time it moves to different states and makes decisions on which action to choose for moving from current state to the next state.
Dalia BaziukaitÄ—
doaj +1 more source
Reinforcement Learning: Stochastic Approximation Algorithms for Markov Decision Processes [PDF]
This article presents a short and concise description of stochastic approximation algorithms in reinforcement learning of Markov decision processes. The algorithms can also be used as a suboptimal method for partially observed Markov decision processes.
arxiv
The Value Functions of Markov Decision Processes [PDF]
We provide a full characterization of the set of value functions of Markov decision processes.
arxiv
Sufficiency of Markov Policies for Continuous-Time Jump Markov Decision Processes [PDF]
This paper extends to Continuous-Time Jump Markov Decision Processes (CTJMDP) the classic result for Markov Decision Processes stating that, for a given initial state distribution, for every policy there is a (randomized) Markov policy, which can be defined in a natural way, such that at each time instance the marginal distributions of state-action ...
arxiv
Markov Decision Processes with Incomplete Information and Semi-Uniform Feller Transition Probabilities [PDF]
This paper deals with control of partially observable discrete-time stochastic systems. It introduces and studies Markov Decision Processes with Incomplete Information and with semi-uniform Feller transition probabilities. The important feature of these models is that their classic reduction to Completely Observable Markov Decision Processes with ...
arxiv
Text Classification: A Sequential Reading Approach
We propose to model the text classification process as a sequential decision process. In this process, an agent learns to classify documents into topics while reading the document sentences sequentially and learns to stop as soon as enough information ...
Denoyer, Ludovic+2 more
core +3 more sources
Underwater chemical plume tracing based on partially observable Markov decision process
Chemical plume tracing based on autonomous underwater vehicle uses chemical as a guidance to navigate and search in the unknown environments. To solve the key issue of tracing and locating the source, this article proposes a path-planning strategy based ...
Jiu Hai-Feng+3 more
doaj +1 more source
A Notation for Markov Decision Processes [PDF]
This paper specifies a notation for Markov decision processes.
arxiv
On the detection of Markov decision processes
We study the detection problem for a finite set of Markov decision processes (MDPs) where the MDPs have the same state and action spaces but possibly different probabilistic transition functions. Any one of these MDPs could be the model for some underlying controlled stochastic process, but it is unknown a priori which MDP is the ground truth.
Duan, Xiaoming+4 more
openaire +2 more sources
Optimal policies for discrete time risk processes with a Markov chain investment model [PDF]
We consider a discrete risk process modelled by a Markov Decision Process. The surplus could be invested in stock market assets. We adopt a realistic point of view and we let the investment return process to be statistically dependent over time.
Diasparra, Maikol, Romera, Rosario
core +1 more source