Computing a Mechanism for a Bayesian and Partially Observable Markov Approach
The design of incentive-compatible mechanisms for a certain class of finite Bayesian partially observable Markov games is proposed using a dynamic framework.
Clempner Julio B., Poznyak Alexander S.
doaj +1 more source
A Convex Programming Approach for Discrete-Time Markov Decision Processes under the Expected Total Reward Criterion [PDF]
In this work, we study discrete-time Markov decision processes (MDPs) under constraints with Borel state and action spaces and where all the performance functions have the same form of the expected total reward (ETR) criterion over the infinite time horizon. One of our objective is to propose a convex programming formulation for this type of MDPs.
Dufour, François, Genadot, Alexandre
openaire +3 more sources
Learn Quasi-Stationary Distributions of Finite State Markov Chain
We propose a reinforcement learning (RL) approach to compute the expression of quasi-stationary distribution. Based on the fixed-point formulation of quasi-stationary distribution, we minimize the KL-divergence of two Markovian path distributions induced
Zhiqiang Cai, Ling Lin, Xiang Zhou
doaj +1 more source
A Self-Driving Decision Making With Reachable Path Analysis and Interaction-Aware Speed Profiling
This paper proposes a behavior planning algorithm for self-driving vehicles to handle lane keeping, speed control considering inter-vehicle space, and collision avoidance under uncertainty. The behavior planning approach is structured as a hierarchically
Yuho Song, Sangwon Han, Kunsoo Huh
doaj +1 more source
Using the probabilistic evaluation tool for the analytical solution of large Markov models [PDF]
Stochastic Petri net-based Markov modeling is a potentially very powerful and generic approach for evaluating the performance and dependability of many different systems, such as computer systems, communication networks, manufacturing systems, etc.
Haverkort, Boudewijn R. +1 more
core +4 more sources
A versatile infinite-state Markov reward model to study bottlenecks in 2-hop ad hoc networks [PDF]
In a 2-hop IEEE 801.11-based wireless LAN, the distributed coordination function (DCF) tends to equally share the available capacity among the contending stations. Recently alternative capacity sharing strategies have been made possible.
Cloth, Lucia +2 more
core +2 more sources
Multiphase until formulas over Markov reward models: An algebraic approach
zbMATH Open Web Interface contents unavailable due to conflicting licenses.
Xu, M. +4 more
openaire +3 more sources
Discrete-time rewards model-checked [PDF]
This paper presents a model-checking approach for analyzing discrete-time Markov reward models. For this purpose, the temporal logic probabilistic CTL is extended with reward constraints.
A. Aziz +14 more
core +2 more sources
A Linear Programming Approach to Markov Reward Error Bounds for Queueing Networks [PDF]
In this paper, we present a numerical framework for constructing bounds on stationary performance measures of random walks in the positive orthant using the Markov reward approach. These bounds are established in terms of stationary performance measures of a perturbed random walk whose stationary distribution is known explicitly.
Xinwei Bai, Jasper Goseling
openaire +2 more sources
Indirect Dynamic Negotiation in the Nash Demand Game
The paper addresses a problem of sequential bilateral bargaining with incomplete information. We proposed a decision model that helps agents to successfully bargain by performing indirect negotiation and learning the opponent’s model ...
Tatiana V. Guy +2 more
doaj +1 more source

