ETH Zurich Fall 2017
-
Updated
Apr 12, 2018 - Python
ETH Zurich Fall 2017
Implementation of the Multi-Armed Bandit where each arm returns continuous numerical rewards. Covers Epsilon-Greedy, UCB1, and Thompson Sampling with detailed explanations.
Add a description, image, and links to the linearucb topic page so that developers can more easily learn about it.
To associate your repository with the linearucb topic, visit your repo's landing page and select "manage topics."