Applied Sciences (Jan 2023)

Tensor Implementation of Monte-Carlo Tree Search for Model-Based Reinforcement Learning

  • Marek Baláž,
  • Peter Tarábek

DOI
https://doi.org/10.3390/app13031406
Journal volume & issue
Vol. 13, no. 3
p. 1406

Abstract

Read online

Monte-Carlo tree search (MCTS) is a widely used heuristic search algorithm. In model-based reinforcement learning, MCTS is often utilized to improve action selection process. However, model-based reinforcement learning methods need to process large number of observations during the training. If MCTS is involved, it is necessary to run one instance of MCTS for each observation in every iteration of training. Therefore, there is a need for efficient method to process multiple instances of MCTS. We propose a MCTS implementation that can process batch of observations in fully parallel fashion on a single GPU using tensor operations. We demonstrate efficiency of the proposed approach on a MuZero reinforcement learning algorithm. Empirical results have shown that our method outperforms other approaches and scale well with increasing number of observations and simulations.

Keywords