This package implements the Monte-Carlo Tree Search algorithm in Julia for solving Markov decision processes (MDPs). The user should define the problem according to the generative interface in POMDPs.jl. Examples of problem definitions can be found in POMDPModels.jl.
There is also a BeliefMCTSSolver that solves a POMDP by converting it to an MDP in the belief space.
Special thanks to Jon Cox for writing the original version of this code.
For reference, see the UCT algorithm in this paper: Kocsis, Levente, and Csaba Szepesvári. "Bandit Based Monte-Carlo planning." European Conference on Machine Learning. Springer, Berlin, Heidelberg, 2006.
In Julia, type,
Documentation can be found on the following site: juliapomdp.github.io/MCTS.jl/latest/
mdp is an MDP defined with the POMDPs.jl interface, the MCTS solver can be used to find an optimized action,
a, for the MDP in state
s as follows:
using POMDPs using POMDPModels # for the SimpleGridWorld problem using MCTS using StaticArrays mdp = SimpleGridWorld() solver = MCTSSolver(n_iterations=50, depth=20, exploration_constant=5.0) planner = solve(solver, mdp) a = action(planner, SA[1,2])
See this notebook for an example of how to visualize the search tree.
See this notebook for examples of customizing solver behavior, specifically the Rollouts section for using heuristic rollout policies.