[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: computer-go: Results on self play vs recorded games



On Wednesday 20 August 2003 10:16, Peter Drake wrote:
> We've got some results for 9x9 Go, trained through
> combinations of 3000 games of self play and the 3205

self-play usually takes at least 1-2 orders of magnitude 
more games than learning from existing games.

Did your learning converge? In my experiments the playing 
strength of the networks (tested against GnuGo) is still 
increasing even if the final and temporal errors do not 
change much any more. So it is useful to check the playing 
strength at regular intervals during training to decide 
when to stop the training.

What exploration strategy did you use?

Also 4 games is surely not enough statistics. In my tests I 
usually play 100 games on 9x9 to estimate the playing 
strength and the standard error of the score after sorting 
out duplicates and calculating the variance of the results 
is still about 5 points. Probably even more, because of 
partial correlataions in the games.

- Markus