Results 81 to 90 of about 33,931 (272)
Similarity and Consistency in Algorithm‐Guided Exploration
ABSTRACT Algorithmic advice has the potential to significantly improve human decision‐making, especially in dynamic and complex tasks that require a balance between exploration and exploitation. This study examines conditions under which individuals are willing to accept advice from algorithms in such scenarios, focusing on the interaction between ...
Ludwig Danwitz +6 more
wiley +1 more source
Photonic Ising Machine Using Parallel Bandit Architecture Based on Laser Chaos
Photonic computing has emerged as a transformative technology for overcoming the fundamental limitations of semiconductor technology. Among recent advancements, the laser chaos decision maker has been introduced for ultrafast decision making by ...
Ryota Yasudo +2 more
doaj +1 more source
Causally Abstracted Multi-armed Bandits
8 pages, 3 figures (main article); 20 pages, 10 figures (appendix); 40th Conference on Uncertainty in Artificial Intelligence (UAI)
Zennaro, FM +6 more
openaire +4 more sources
Reinforcement Learning in Modern Biostatistics: Constructing Optimal Adaptive Interventions
Summary In recent years, reinforcement learning (RL) has acquired a prominent position in health‐related sequential decision‐making problems, gaining traction as a valuable tool for delivering adaptive interventions (AIs). However, in part due to a poor synergy between the methodological and the applied communities, its real‐life application is still ...
Nina Deliu +2 more
wiley +1 more source
Application of Multi-Armed Bandit Algorithm in Quantitative Finance [PDF]
The volatility and diversity of financial markets make it challenging for a single portfolio achieve better returns, therefore, adjustable portfolios based on the risk tolerance of clients are highly demanded.
Chen Chengxun +3 more
doaj +1 more source
Treatment effect optimisation in dynamic environments
Applying causal methods to fields such as healthcare, marketing, and economics receives increasing interest. In particular, optimising the individual-treatment-effect – often referred to as uplift modelling – has peaked in areas such as precision ...
Berrevoets Jeroen +2 more
doaj +1 more source
Budgeted Combinatorial Multi-Armed Bandits
15 pages, 4 figures. To be published in AAMAS 2022.
Das, Debojit, Jain, Shweta, Gujar, Sujit
openaire +2 more sources
Pure Exploration for Multi-Armed Bandit Problems [PDF]
We consider the framework of stochastic multi-armed bandit problems and study the possibilities and limitations of forecasters that perform an on-line exploration of the arms.
Bubeck, Sébastien +2 more
core +4 more sources
Enhancing Evolutionary Conversion Rate Optimization via Multi-armed Bandit Algorithms
Conversion rate optimization means designing web interfaces such that more visitors perform a desired action (such as register or purchase) on the site. One promising approach, implemented in Sentient Ascend, is to optimize the design using evolutionary ...
Miikkulainen, Risto, Qiu, Xin
core +1 more source
We consider a variant of the classic multi-armed bandit problem where the expected reward of each arm is a function of an unknown parameter. The arms are divided into different groups, each of which has a common parameter. Therefore, when the player selects an arm at each time slot, information of other arms in the same group is also revealed.
Wang, Zhiyang, Zhou, Ruida, Shen, Cong
openaire +2 more sources

