Multi-Gear Bandits, Partial Conservation Laws, and Indexability
This paper considers what we propose to call <i>multi-gear bandits</i>, which are Markov decision processes modeling a generic dynamic and stochastic <i>project</i> fueled by a single resource and which admit multiple actions representing gears of operation naturally ordered...
Main Author: | José Niño-Mora |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-07-01
|
Series: | Mathematics |
Subjects: | |
Online Access: | https://www.mdpi.com/2227-7390/10/14/2497 |
Similar Items
-
A Fast-Pivoting Algorithm for Whittle’s Restless Bandit Index
by: José Niño-Mora
Published: (2020-12-01) -
Markovian Restless Bandits and Index Policies: A Review
by: José Niño-Mora
Published: (2023-03-01) -
Fast Two-Stage Computation of an Index Policy for Multi-Armed Bandits with Setup Delays
by: José Niño-Mora
Published: (2020-12-01) -
Design of Multi-Armed Bandit-Based Routing for in-Network Caching
by: Gen Tabei, et al.
Published: (2023-01-01) -
Signal detection models as contextual bandits
by: Thomas N. Sherratt, et al.
Published: (2023-06-01)