π° Bandit AlgorithmsMulti-Armed Bandits, Thompson Sampling, UCB, Exploration Strategy, Online Learning