I was thinking the same thing. You can easily equip the value network with
several outputs, corresponding to several settings of komi, then train as
usual.
The issue with Japanese rules is easily solved by refusing to play under
ridiculous rules. Yes, I do have strong opinions. :)
Álvaro.
On
I guess that 1 point in such a game matters to the evaluation function.
Pretty fascinating. Can you not train for the two different rulesets and
just pick which at the beginning? Ignoring Chinese versus Japanese, just
training on komi? Or is the problem of Japanese rules the whole issue? (I.e
not
How does Zen know it's playing a Japanese rule game? Can it be set to play
a Chinese rule game and hope to converge at the end?
On Mar 21, 2017 8:03 AM, "Hideki Kato" wrote:
> The value network has been trained with Chinese rules and 7.5
> pts komi. Using this for
Why can't you reuse the same self played games but score them with a
different komi value ? The policy network does not use the komi to choose
its moves so it should make no difference.
> On 21/03/2017 21:08, David Ongaro wrote:
>>> But how would you fix it? Isn't that you'd need to retrain
On 21/03/2017 21:08, David Ongaro wrote:
>> But how would you fix it? Isn't that you'd need to retrain your value
>> network from the scratch?
>
> I would think so as well. But I some months ago I already made a
> proposal in this list to mitigate that problem: instead of training a
> different
Hi,
now we see how clever the DeepMind team was (and likely still is).
In both matches (against Fan Hui and Lee Sedol) Chinese rules
were applied.
Some years ago I performed experiments with Monte Carlo search
in special non-zero sum games (with
Hideki,
Using this for Japanese and 6.5, there will be some
>> error in close games. We knew this issue and thought such
>> chances would be so small that postponed correcting (not so
>> easy).
>>
>
But how would you fix it? Isn't that you'd need to retrain your value
network from the scratch?
On Tue, Mar 21, 2017 at 10:48 AM, Hideki Kato
wrote:
> The value network has been trained with Chinese rules and 7.5
> pts komi. Using this for Japanese and 6.5, there will be some
> error in close games. We knew this issue and thought such
> chances would be so small
The value network has been trained with Chinese rules and 7.5
pts komi. Using this for Japanese and 6.5, there will be some
error in close games. We knew this issue and thought such
chances would be so small that postponed correcting (not so
easy).
Best,
Hideki
Pawe Morawiecki:
On 21 March 2017 at 03:45, "Ingo Althöfer" <3-hirn-ver...@gmx.de> wrote:
> * Will AlphaGo accept the challenge by FineArt?
> (or Google the one by 10Cent, respectively ?)
>
> * Will DeepZenGo win the World Championships, starting right now?
>
> * Will some Robin Good rise from the underwoods to
>
> * Which of the currently three top bots will show up in the
> European Go Congress in Oberhof in July/August 2017?
just set up one of the top open source bots:
on moderate hardware
ray: http://www.dragongoserver.net/userinfo.php?uid=97868
and if this is too strong for europe
oakfoam:
11 matches
Mail list logo