Performance of Multi-Armed Bandit Algorithms in Dynamic vs. Static Environments: A Comparative Analysis

This paper conducts a comparative analysis of Multi-Armed Bandit (MAB) algorithms, particularly the Upper Confidence Bound (UCB) and Thompson Sampling (TS) algorithms, and focuses on the performance of these algorithms in both static and dynamic environments. Multi-armed bandit algorithms are instru...

Full description

Bibliographic Details
Main Author: Zhao Boxi
Format: Article
Language:English
Published: EDP Sciences 2025-01-01
Series:ITM Web of Conferences
Online Access:https://www.itm-conferences.org/articles/itmconf/pdf/2025/04/itmconf_iwadi2024_01016.pdf