Re: [Computer-go] Value network that doesn't want to learn.

2017-06-23 Thread Brian Sheppard via Computer-go
>... my value network was trained to tell me the game is balanced at the >beginning... :-) The best training policy is to select positions that correct errors. I used the policies below to train a backgammon NN. Together, they reduced the expected loss of the network by 50% (cut the error

Re: [Computer-go] Value network that doesn't want to learn.

2017-06-23 Thread Vincent Richard
Finally found the problem. In the end, it was as stupid as expected: When I pick a game for the batch creation I select randomly a limited number of moves inside the game. In the case of the value network I use like 8-16 moves to not overfit the data (I can't take 1 or then the I/O operations

Re: [Computer-go] Value network that doesn't want to learn.

2017-06-19 Thread Vincent Richard
This is what have been thinking about, yet unable to find an error. Currently, I'm working with: - SGF Database: fuseki info Tygem -> http://tygem.fuseki.info/index.php (until recently I was working with games of all level from KGS) - The data is then analyzed by a script which extracts all

Re: [Computer-go] Value network that doesn't want to learn.

2017-06-19 Thread Gian-Carlo Pascutto
On 19-06-17 17:38, Vincent Richard wrote: > During my research, I’ve trained a lot of different networks, first on > 9x9 then on 19x19, and as far as I remember all the nets I’ve worked > with learned quickly (especially during the first batches), except the > value net which has always been

[Computer-go] Value network that doesn't want to learn.

2017-06-19 Thread Vincent Richard
Hello everyone, For my master thesis, I have built an AI that has a strategical approach to the game. It doesn’t play but simply describe the strategy behind all possible move for a given strategy ("enclosing this group", "making life for this group", "saving these stones", etc). My main idea

Re: [Computer-go] Value Network

2016-03-19 Thread Detlef Schmicker
; } } } layers { name: >> "relu10" type: RELU bottom: "conv10" top: "conv10" } >> >> layers { name: "conv11_3x3_128" type: CONVOLUTION blobs_lr: 1. >> blobs_lr: 2. bottom: "conv10" top: "conv11" convolution_param

Re: [Computer-go] Value Network

2016-03-19 Thread Hiroshi Yamashita
uang" <ajahu...@google.com> To: <computer-go@computer-go.org> Sent: Saturday, March 19, 2016 10:25 PM Subject: Re: [Computer-go] Value Network Good stuff, Hiroshi. Looks like I don't need to answer the questions regarding value network. :) Aja ___

Re: [Computer-go] Value Network

2016-03-19 Thread Hiroshi Yamashita
utput: 1 kernel_size: 1 pad: 0 weight_filler { type: "xavier" } bias_filler { type: "constant" } } } layers { name: "relu12" type: RELU bottom: "conv12" top: "conv12" } layers { name: "fc13" type: INNER_PRODUCT

Re: [Computer-go] Value Network

2016-03-19 Thread Aja Huang
weight_filler { > type: "xavier" >} >bias_filler { > type: "constant" >} > } > } > layers { > name: "relu10" > type: RELU > bottom: "conv10" > top: "conv10" > } > > layers { > na

Re: [Computer-go] Value Network

2016-03-19 Thread Detlef Schmicker
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 What are you using for loss? this: layers { name: "loss4" type: EUCLIDEAN_LOSS loss_weight: 2.0 bottom: "vvv" bottom: "pool2" top: "accloss4" } ? Am 04.03.2016 um 16:23 schrieb Hiroshi Yamashita: > Hi, > > I tried to make Value

Re: [Computer-go] Value Network

2016-03-04 Thread Detlef Schmicker
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Hi, thanks a lot for sharing! I try a slightly different approach at the moment: I use a combined policy / value network (adding 3-5 layers with about 16 filters at the end of the policy network for the value network to avoid overfitting) and I use

[Computer-go] Value Network

2016-03-04 Thread Hiroshi Yamashita
Hi, I tried to make Value network. "Policy network + Value network" vs "Policy network" Winrate Wins/Games 70.7%322 / 455,1000 playouts/move 76.6%141 / 184, 1 playouts/move It seems more playouts, more Value network is effetctive. Games is not enough though. Search is