Hedging the Drift: Learning to Optimize Under Nonstationarity
<jats:p> We introduce data-driven decision-making algorithms that achieve state-of-the-art dynamic regret bounds for a collection of nonstationary stochastic bandit settings. These settings capture applications such as advertisement allocation, dynamic pricing, and traffic network routing in c...
Main Authors: | Cheung, Wang Chi, Simchi-Levi, David, Zhu, Ruihao |
---|---|
Other Authors: | Massachusetts Institute of Technology. Department of Civil and Environmental Engineering |
Format: | Article |
Language: | English |
Published: |
Institute for Operations Research and the Management Sciences (INFORMS)
2023
|
Online Access: | https://hdl.handle.net/1721.1/148652 |
Similar Items
-
Learning to Optimize Under Non-Stationarity
by: Cheung, Wang Chi, et al.
Published: (2021) -
Reinforcement Learning for Non-Stationary Markov Decision Processes: The Blessing of (More) Optimism
by: Cheung, Wang Chi, et al.
Published: (2021) -
Meta Dynamic Pricing: Transfer Learning Across Experiments
by: Bastani, Hamsa, et al.
Published: (2023) -
Near-Optimal Model-Free Reinforcement Learning in Non-Stationary Episodic MDPs
by: Mao, Weichao, et al.
Published: (2023) -
Adjusting the neuroimaging statistical inferences for nonstationarity.
by: Salimi-Khorshidi, G, et al.
Published: (2009)