This project contains my work on the Multi-Armed Bandit (MAB) problem and variations and extensions thereof. This work was carried out as part of an Independent Study in Computer Science during university. This problem in encountered within probability theory, statistics, and machine learning (specifically, Reinforcement Learning). It is well-regarded as an improvement upon the widely adopted A/B Testing approach, due to the model's dynamic learning and incorporation of outcomes and feedback. A central concept within MABs is that of the tradeoff between exploration (learning about options) and exploitation (maximizing the current best option).
-
Notifications
You must be signed in to change notification settings - Fork 0
krishnarupa1008/Multi-Armed-Bandits-Combinatorial-Multi-Armed-Bandits
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published