- Example Tic-Tac-Toe
- Example: Solving Tic-Tac-Toe with And-Or-Tree Search. Here the opponent is seen as part of the environment, i.e., each action by the player is followed by an unknown action of the opponent which, from the viewpoint of the player makes the outcomes of actions nondeterministic.
- Example: Solving Tic-Tac-Toe with Minimax Search and Alpha-Beta Pruning
- Example: Solving Tic-Tac-Toe with Heuristic Alpha-Beta Tree Search
- Example: Solving Tic-Tac-Toe with Pure Monte Carlo Search
- Example: Solving Tic-Tac-Toe with Pure Monte Carlo Search + UCB1 Selection Policy
- Assignment: Adversarial Search: Playing Connect 4
- Assignment: Adversarial Search: Playing "Mean" Connect 4
- Assignment: Adversarial Search: Playing Dots and Boxes
- Example: Learn to Score a Tic-Tac-Toe Board by Example uses a neural network trained on self-play data to lean an evaluation function that can be used as a heuristic in Heuristic Minimax Search or as a playout policy for Monte Carlo Search.
- Example: Model-free Reinforcement Learning to Play Tic-Tac-Toe with Q-Learning.
- Example: Model-based Reinforcement Learning to Play Tic-Tac-Toe with Value Iteration.
All code and documents in this repository is provided under Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0) License
