Performance of Multi-Armed Bandit Algorithms in Dynamic vs. Static Environments: A Comparative Analysis
This paper conducts a comparative analysis of Multi-Armed Bandit (MAB) algorithms, particularly the Upper Confidence Bound (UCB) and Thompson Sampling (TS) algorithms, and focuses on the performance of these algorithms in both static and dynamic environments. Multi-armed bandit algorithms are instru...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
EDP Sciences
2025-01-01
|
Series: | ITM Web of Conferences |
Online Access: | https://www.itm-conferences.org/articles/itmconf/pdf/2025/04/itmconf_iwadi2024_01016.pdf |