This is a quick check of my understanding of the network architecture. Let's count the number of parameters in the model: * convolutional block: (17*9+1)*256 + 2*256 [ 17 = number of input channels 9 = size of the 3x3 convolution window 1 = bias (I am not sure this is needed if you are going to do batch normalization immediately after) 256 = number of output channels 2 = mean and standard deviation of the output of the batch normalization 256 = number of channels in the batch normalization ] * residual block: (256*9+1)*256 + 2*256 + (256*9+1)*256 + 2*256 * policy head: (256*1+1)*2 + 2*2 + (2*361+1)*362 * value head: (256*1+1)*1 + 2*1 + (1*361+1)*256 + (256+1)*1
Summing it all up, I get 22,837,864 parameters for the 20-block network and 46,461,544 parameters for the 40-block network. Does this seem correct? Álvaro. On Thu, Oct 19, 2017 at 6:17 AM, Petr Baudis <pa...@ucw.cz> wrote: > On Wed, Oct 18, 2017 at 04:29:47PM -0700, David Doshay wrote: > > I saw my first AlphaGo Zero joke today: > > > > After a few more months of self-play the games might look like this: > > > > AlphaGo Zero Black - move 1 > > AlphaGo Zero White - resigns > > ...which is exactly what my quick attempt to reproduce AlphaGo Zero > yesterday converged to overnight. ;-) But I'm afraid it's because of > a bug, not wisdom... > > Petr Baudis > _______________________________________________ > Computer-go mailing list > Computer-go@computer-go.org > http://computer-go.org/mailman/listinfo/computer-go >
_______________________________________________ Computer-go mailing list Computer-go@computer-go.org http://computer-go.org/mailman/listinfo/computer-go