Ray icon indicating copy to clipboard operation
Ray copied to clipboard

Suggestion for solving endgame problem

Open j8580 opened this issue 8 years ago • 17 comments

Ray or many go ai has endgame problem except alphago. I suggest some idea. Make new command for endgame. In particular part of game, with endgame command Ray use only counting result search. Best candidate is selected for maximum counting result. Or ray's neural network plus maximum counting search will be good choice for good endgame performance. Is it possible?

j8580 avatar Apr 21 '17 10:04 j8580

I think it's open problem. Please try it and send PR or write the paper.

zakki avatar Apr 21 '17 13:04 zakki

Some idea.

  1. Simultaneously use multi neural network.(1 2 3 4 5...) Ex) if player is qhite set for multi komi.
  • 1 NN 6.5
  • 2 Nn 5.5
  • 3 NN 4.5
  • 4 Nn 3.5 and so on
  • 5 ..... If player is black vice versa
  • 1 Nn 6.5
  • 2 Nn 7.5
  • 3 NN 8.5 and so on

In early stage or middle satge of game it,s not important. But in end game there r huge differences in win percentage. Ray can select higer win ratio move in multi nn . And multi nn option works among ex) 40 ~60 winning or losing ratio. Ray van turn on or turn off multi nn option. Is this possible ?

j8580 avatar Apr 25 '17 11:04 j8580

Ray can turn on or turn off multi nn option with some algorithm.

j8580 avatar Apr 25 '17 11:04 j8580

If simultaneous nn is impossible, then ray use sequential nn search.

j8580 avatar Apr 25 '17 11:04 j8580

CNTK evaluation library can load multiple model. Also you can merge multiple models into single model, and switch them using BrainScript. https://github.com/Microsoft/CNTK/wiki/CloneFunction https://github.com/Microsoft/CNTK/wiki/If-Operation

zakki avatar Apr 25 '17 11:04 zakki

Well i am not developer.that was just my tip of endgame problem. I thnik u can do it . Ur genius. But I am doing some tests my idea. And if it is useful then I will inform more details. Then u can adjust it to ray.

j8580 avatar Apr 25 '17 12:04 j8580

Ray is enough strong almost professional level except endgame. If u can solve ray's endgame problem ray will be first professional level go game in personal pc level.

j8580 avatar Apr 25 '17 12:04 j8580

It may be possible, but approximately it requires 10TB SSD to store data, 2 week to generate training data, 2 month to training. It is too large for me.

zakki avatar Apr 25 '17 12:04 zakki

Now Ray uses only games that human plays to train. Data generation with RL policy and random play like AlphaGo may improve endgame. But that requires enormous resource too. 😱

zakki avatar Apr 25 '17 12:04 zakki

@j8580 CGI published "Multi-Labelled Value Networks for Computer Go" https://arxiv.org/abs/1705.10701 I haven't understood detail, but it looks similar to your idea in some aspect.

zakki avatar May 31 '17 09:05 zakki

I will check papers.

j8580 avatar May 31 '17 09:05 j8580

Lol their dynamic komi concept is my ideas ^^ I have tried many times in manually with Ray. It was really useful. And still testing with ray. I will read it in details.i can not sure I can get their exact ideas. Give me a some time.

j8580 avatar May 31 '17 10:05 j8580

sorry..i couldnt understand the exact methods.

just I tried manually dynamic komi change during play.

komi change only effects on MCTS not neural network. only MCTS value change is also useful.

but I can also see neural network value effect when dynamic komi changes during play.

can u add some code? (if komi change during play..then nn value also change)

I feel komi change is useful.

and end game...max policy candidate is more accurate than ray's best candidat in many times.

really sorry for not helping u.

Still I hope u improve and solve Ray's strength..cheer up~

j8580 avatar Jun 05 '17 03:06 j8580

https://github.com/zakki/Ray/blob/nn/src/UctSearch.cpp#L2539 Is what you mean replacing komi to dynamic_komi here?

zakki avatar Jun 05 '17 03:06 zakki

i change komi during play by setkomi command or gogui komi set.

j8580 avatar Jun 05 '17 03:06 j8580

example..if white..at start..komi is 4.5 ...then nn winning ration is over 55 then white komi 3.5 ..then if nn winning ratio is over 60 then white komi is 1.5 and so on. and winng percentage is over 80 then ray use max policy endgame... i don know optimal value. they said optimal values in papers. but i cant understand their papers exactly.

i have played with pro go players with ray.. now ray's winning ratio is 30-40% vs average professional players with this manual technique...but just this is manual technique..i don know ray can this automatically. and ray has still some bugs in big life-death situation. especially throw-in bugs.

it's just examples..and now komi change only effect on MCTS winning percentage not nn value. but still i feel it's useful.

j8580 avatar Jun 05 '17 03:06 j8580

at endgame..by manually i choose the one among best 3 max policy candidate not ray's best candidate. if ray's best candidate is same like max policy candidate, then that is best at endgame.. i think ruling out too low policy value candidate is more stable. i am tygem 5d in real life.. so i can know approximately endgame situation. then i can choose manual candidate with Ray. but how can Ray know endgame situation?..can u do? when Ray is winning at endgame part(example ray's winning ratio is over 80) , in order to reduce Ray's endgame mistakes, i suggest Ray choose the one among top 3 high policy value candidate. surely that is satble in my experience. by this I've never experienced a reversal defeat recently at endgame part when Ray is winning.

j8580 avatar Jun 07 '17 02:06 j8580