Results 311 to 320 of about 1,517,214 (350)
Some of the next articles are maybe not open access.
A Partially Observable Markov Decision Process Approach to Residential Home Energy Management
IEEE Transactions on Smart Grid, 2018Real-time pricing (RTP) is a utility-offered dynamic pricing program to incentivize customers to make changes in their energy usage. A home energy management system (HEMS) automates the energy usage in a smart home in response to utility pricing signals.
T. Hansen+4 more
semanticscholar +1 more source
Reinforcement Learning to Rank with Markov Decision Process
Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 2017One of the central issues in learning to rank for information retrieval is to develop algorithms that construct ranking models by directly optimizing evaluation measures such as normalized discounted cumulative gain~(ND CG).
Zheng Wei+4 more
semanticscholar +1 more source
Adapting Markov Decision Process for Search Result Diversification
Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 2017In this paper we address the issue of learning diverse ranking models for search result diversification. Typical methods treat the problem of constructing a diverse ranking as a process of sequential document selection.
Long Xia+5 more
semanticscholar +1 more source
Intelligent decision making for overtaking maneuver using mixed observable Markov decision process
Journal of Intelligent Transportation Systems / Taylor & Francis, 2018Overtaking maneuver is one of the most dangerous scenarios for road vehicles especially in two-way roads. In this article, we propose a new formulation for the problem of overtaking in two-way roads using the tools from the Mixed Observable Markov ...
Volkan Sezer
semanticscholar +1 more source
Partially Observable Markov Decision Processes and Robotics
Annu. Rev. Control. Robotics Auton. Syst., 2022Planning under uncertainty is critical to robotics. The partially observable Markov decision process (POMDP) is a mathematical framework for such planning problems.
H. Kurniawati
semanticscholar +1 more source
Markov Decision Processes [PDF]
This chapter introduces sequential decision problems, in particular Markov decision processes (MDPs). A formal definition of an MDP is given, and the two most common solution techniques are described: value iteration and policy iteration. Then, factored MDPs are described, which provide a representation based on graphical models to solve very large ...
openaire +1 more source
A Markov Decision Process-based service migration procedure for follow me cloud
2014 IEEE International Conference on Communications (ICC), 2014The Follow-Me Cloud (FMC) concept enables service mobility across federated data centers (DCs). Following the mobility of a mobile user, the service located in a given DC is migrated each time an optimal DC is detected.
A. Ksentini, T. Taleb, Min Chen
semanticscholar +1 more source
Markov Decision Process-Based Distributed Conflict Resolution for Drone Air Traffic Management
, 2017Ensuring safety and providing timely conflict alerts to small unmanned aircraft, commonly known as drones, is important to their integration into civil airspace. This paper proposes a short-term conflict avoidance algorithm for an automated low-altitude,
H. Ong, Mykel J. Kochenderfer
semanticscholar +1 more source
2021
As discussed in Chapter 1, reinforcement learning involves sequential decision-making. In this chapter, we will formalize the notion of using stochastic processes under the branch of probability that models sequential decision-making behavior. While most of the problems we study in reinforcement learning are modeled as Markov decision processes (MDP ...
openaire +2 more sources
As discussed in Chapter 1, reinforcement learning involves sequential decision-making. In this chapter, we will formalize the notion of using stochastic processes under the branch of probability that models sequential decision-making behavior. While most of the problems we study in reinforcement learning are modeled as Markov decision processes (MDP ...
openaire +2 more sources
Semi-Markov Decision Process With Partial Information for Maintenance Decisions
IEEE Transactions on Reliability, 2014A critical factor that prevents optimal scheduling of maintenance interventions is the uncertainty regarding the current condition of the asset under consideration, as well as the rate at which deterioration takes place.
R. Srinivasan, A. Parlikad
semanticscholar +1 more source