You are totally right it is not the same curves. according to the reddit
post.
So I was totally wrong
> On 27-10-17 10:15, Xavier Combelle wrote:
>> Maybe I'm wrong but both curves for alphago zero looks pretty similar
>> except than the figure 3 is the zoom in of figure 6
> The blue curve in
OK I will reread it attentively
Le 27/10/2017 à 19:19, Hideki Kato a écrit :
> Please read _through_ the paper sequentially.
> #I don't have enough skill to describe the reason because
> it's not a technical but language issue.
>
> Hideki
>
>> I don't understand which element makes you say that
Please read _through_ the paper sequentially.
#I don't have enough skill to describe the reason because
it's not a technical but language issue.
Hideki
>I don't understand which element makes you say that
>section 2 and 3 are all for a 20 block instance
>
>
>Le 27/10/2017 E01:49, Hideki Kato a
On 27-10-17 10:15, Xavier Combelle wrote:
> Maybe I'm wrong but both curves for alphago zero looks pretty similar
> except than the figure 3 is the zoom in of figure 6
The blue curve in figure 3 is flat at around 60 hours (2.5 days). In
figure 6, at 2.5 days the line is near vertical. So it is
Maybe I'm wrong but both curves for alphago zero looks pretty similar
except than the figure 3 is the zoom in of figure 6
Le 27 oct. 2017 04:31, "Gian-Carlo Pascutto" a écrit :
> Figure 6 has the same graph as Figure 3 but for 40 blocks. You can compare
> the Elo.
>
> On Thu,
The 40 block version (2nd instance) first appeared in
Section 4 in the paper. Section 2 and 3 are all for the 1st
instance.
Hideki
Xavier Combelle: <39a79a0e-7c7d-2a01-a2ae-573cda8b1...@gmail.com>:
>Unless I mistake figure 3 shows the plot of supervised learning to
>reinforcement learning,
Figure 6 has the same graph as Figure 3 but for 40 blocks. You can compare
the Elo.
On Thu, Oct 26, 2017, 23:35 Xavier Combelle
wrote:
> Unless I mistake figure 3 shows the plot of supervised learning to
> reinforcement learning, not 20 bloc/40 block
>
> For searching
com>:
>>>>>> How is it a fair comparison if there is only 3 days of training
>>>>>> for
>>> Zero ?
>>>>>> Master had longer training no ? Moreover, Zero has bootstrap
>>>>>> problem
>>>>>> because at the op
Unless I mistake figure 3 shows the plot of supervised learning to
reinforcement learning, not 20 bloc/40 block
For searching mention of the 20 blocks I search for 20 in the whole
paper and did not found any other mention
than of the kifu thing.
Le 26/10/2017 à 15:10, Gian-Carlo Pascutto a
t;> Zero ?
>>>>>> Master had longer training no ? Moreover, Zero has bootstrap problem
>>>>>> because at the opposite of Master it don't learn from expert games
>>>>>> which means that it is likely to be we
On 26-10-17 10:55, Xavier Combelle wrote:
> It is just wild guesses based on reasonable arguments but without
> evidence.
David Silver said they used 40 layers for AlphaGo Master. That's more
evidence than there is for the opposite argument that you are trying to
make. The paper certainly
ter it don't learn from expert games
>>>>> which means that it is likely to be weaker with little training.
>>>>> Le 24/10/2017 à 20:20, Hideki Kato a écrit :
>>>>>> David Silver told Master used 40 layers network in May.
>&g
> because at the opposite of Master it don't learn from expert games
>>>> which means that it is likely to be weaker with little training.
>>>> Le 24/10/2017 à 20:20, Hideki Kato a écrit :
>>>>> David Silver told
er told Master used 40 layers network in May.
>>>> According to new paper, Master used the same architecture
>>>> as Zero. So, Master used 20 blocks ResNet.
>>>> The first instance of Zero, 20 blocks ResNet version, is
>>>> weaker than Master (aft
aper, Master used the same architecture
>>> as Zero. So, Master used 20 blocks ResNet.
>>> The first instance of Zero, 20 blocks ResNet version, is
>>> weaker than Master (after 3 days training). So, with the
>>> same layers (a fair comparison) Zero is weaker than
&g
locks ResNet version, is
>> weaker than Master (after 3 days training). So, with the
>> same layers (a fair comparison) Zero is weaker than
>> Master.
>>
>> Hideki
>
>
>___
>Computer-go mailing list
irst instance of Zero, 20 blocks ResNet version, is
> > weaker than Master (after 3 days training). So, with the
> > same layers (a fair comparison) Zero is weaker than
> > Master.
> >
> > Hideki
>
>
> ___
er 3 days training). So, with the
> same layers (a fair comparison) Zero is weaker than
> Master.
>
> Hideki
_______
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go
David Silver told Master used 40 layers network in May.
According to new paper, Master used the same architecture
as Zero. So, Master used 20 blocks ResNet.
The first instance of Zero, 20 blocks ResNet version, is
weaker than Master (after 3 days training). So, with the
same layers (a
19 matches
Mail list logo