gooooloo
gooooloo
+1 for the feature request.
@tellarin Thanks for clarifying. Shall I close the issue or keep it open( in case you want to look into the "should not return anything" thing)? Please feel free to...
Just for your reference, I am sharing tree search between 2 players, see codes here: https://github.com/gooooloo/alpha-zero-in-python/blob/master/src/reversi_zero/agent/player.py But I don't think this makes big difference. Many other settings are much more...
Maybe that is just the nature of Reversi game, see https://en.wikipedia.org/wiki/Computer_Othello, the "Othello 8 x 8" section. That being said, even if enough randomness is promised when traning, it will...
I see. Looking forward to a solution being found~
I have some model reports in `records.md` of this [repo](https://github.com/gooooloo/reversi-alpha-zero/blob/master/records.md)
@apollo-time in my report, ntest:6 means the opponent is NTest with strengh 6, 6/1/3 means 6 wins, 1 draw, 3 lose. By the way, you can find the win/draw game...
@apollo-time > How about policy and value loss? policy loss : 0.4 value loss : 0.4-0.55. Unstable.
@apollo-time > Um...But you use game history, isn't it? Yes. I am actually guessing that, the resnet blocks number in the model should be reduced depends on the historical board...
@richemslie congrats on getting the result. ntest5 is already strong, I feel. > Can you try ntest level 7 again with your latest model I have detailed evaluation metric [here](https://github.com/gooooloo/reversi-alpha-zero/blob/master/records.md)....