We implemented a Monte Carlo Tree Search (MCTS) from scratch and we successfully applied it to Tic-Tac-Toe game.
-
Updated
Jul 9, 2020 - Python
We implemented a Monte Carlo Tree Search (MCTS) from scratch and we successfully applied it to Tic-Tac-Toe game.
Multi Armed Bandits implementation using the Jester Dataset
Source code for Assignment 2 of COMP90051 (Semester 2 2020)
Repository for the course project done as part of CS-747 (Foundations of Intelligent & Learning Agents) course at IIT Bombay in Autumn 2022.
Implementation of Multi-Armed Bandit (MAB) algorithms UCB and Epsilon-Greedy. MAB is a class of problems in reinforcement learning where an agent learns to choose actions from a set of arms, each associated with an unknown reward distribution. UCB and Epsilon-Greedy are popular algorithms for solving MAB problems.
We compare different policies for the checkers game using reinforcement learning algorithms.
R.I.T project
LoRa@FIIT algorithms comparison using jupyter notebooks
Complete Tutorial Guide with Code for learning ML
CS70 Homework and Discussion Solutions
Train and test your IA's using these samples in the machine learning field.
Bandit algorithms in OCaml
This is a sample code written in R that compares Thompson Sampling and UCB for three available arms sampled from a bernoulli distribution.
Add a description, image, and links to the ucb topic page so that developers can more easily learn about it.
To associate your repository with the ucb topic, visit your repo's landing page and select "manage topics."