AlphaZero.jl
AlphaZero.jl copied to clipboard
question on parallel tree search
AGZ uses virtual loss to make the tree search threads explore different paths. What is the strategy used in AlphaZero.jl?
AlphaZero.jl does not have multiple workers explore the same MCTS tree and therefore there is no need for a virtual loss. Instead, many games are simulated in parallel during self-play and network inference requests are batched across those.
We probably still want a mechanism for matches to facilitate a large number of simulations.
This is true and as a matter of fact, matches would be quite slow with the current implementation. This is on my TODO list though.
looking forward to the new release