I think you misunderstood what I wrote,
if perfect play on 9x9 is 6000 Elo, then if the value function is 3000
Elo and MC eval is 2000 Elo with 1 second thinking time then it might
be that the combination of a value function and MC eval ends up being
2700 Elo. It could also be that it ends up at 3200 Elo.
I personally believe MC eval alone can be very strong so it might be
that the capacity of a neural network is not enough to replace MC eval.
When I wrote "converge to perfection" it was not to claim that the Zero
approach reaches perfection, just that it get stronger over time.
The interesting question is if old school MC evaluation can fill up the
knowledge gaps of the value function.
For my own Odin project I am not working on the MC evaluation currently,
since only when I a have final neural network solution can I see which
weaknesses I need to fix.
On 2018-03-05 21:12, Hideki Kato wrote:
DCNNs are not magic but just non-linear continuous function
approximators with finite freedom and we can provide up to
10^8 samples (board positions) in practice.
Why do most people believe VN can approximate (perfect or
near perfect) value function? What do they estimate the
complexity of the value function for 19x19 Go?
My guess is that there is some kind of threshold depending on the
relative strength of MC eval and the value function of the NN.
If the value function is stronger than MC eval I would guess MCEval
turns into a bad noisy feature with little benefit.
Depending on how strong MC eval is this threshold is probably very
different between engines. Also i can imagine that NN value function
have some gaping holes in its knowledge that even simple MC eval can
patch up. Probably true for supervised learning where training data
probably has a lot of holes since bad moves are not in the data.
The Zero approach is different because it should converge to
in the limit, thus overcome any weaknesses of the value function early
on. At least in theory.
On 2018-03-05 14:04, Gian-Carlo Pascutto wrote:
On 5/03/2018 12:28, valky...@phmp.se wrote:
Remi twittered more details here (see the discussion with gghideki:
Thank you. So Remi gave up on rollouts as well. Interesting
of opinion" there with Zen.
Last time I tested this in regular Leela, playouts were beneficial,
this was before combined value+policy nets and much more training
was available. I do not know what the current status would be.
Computer-go mailing list
Computer-go mailing list