I'm all for a learning policy, if you can figure out how to do it. :-)

Peter Drake
http://www.lclark.edu/~drake/



On Jan 25, 2011, at 11:31 AM, Aja wrote:

Hi Professor Drake,

I will try with more playouts. Thanks for your reminding.

I give an example to show my view: default policy should also be included to learn. I suppose: if there are several decisive life-and- death or semeai in a position, the tree search cannot go to/clarify every one of them.

In this example, Black's L2 and L4 will cause White to play L3 to capture by default policy (it's completely bad). Then Black may learn quickly by "last good play" to atari immediately and kill White's whole group to win. The problem is, White is not able to learn the correct answer H1 or H2 because it is fixed in default policy.

In the playouts, the configuration of such a big semeai might be very similar. Such evaluation bias is exactly an issue that we can fix by learning. By considering probability, I can fix this problem by increasing the probability of the "last good reply" H1 or H2, without tree's aid.

Every program's implementation of the playout is more or less different. But I think excluding default policy from the learning might limit the full power of "last good reply".

Aja

----- Original Message -----
From: Peter Drake
To: computer-go@dvandva.org
Sent: Wednesday, January 26, 2011 2:27 AM
Subject: Re: [Computer-go] The heuristic "last good reply"

On Jan 25, 2011, at 10:19 AM, Aja wrote:

Dear all,

Today I have tried Professor Drake's "last good reply" in Erica. So far, I got at most 20-30 elo from it.

I tested by self-play, with 3000 playouts/move on 19x19. The amount of playouts might be too few, but I would like to test more playouts IF the playing strength is not weaker with 3000 playouts.

Yes -- the smallest experiments in the paper were with 8k playouts per move. There may not be time to fill up the reply tables with only 3k.

From this preliminary experiments with 3000 playouts, I have some observations:

1. In Erica, it's better to consider probability for this heuristic.

2. In Prof. Drake's implementation, there is a weakness in learning. I think the main problem is that for a reply which is deterministically played by default policy, there is no room to learn a new reply. For example, if "save by capture" produces a lost game, then in the next simulation, it will still play "save by capture" by default policy. If I am wrong in this point, I am glad to be corrected by anyone.

This is true, but only if the previous move (or previous two moves) come up again in exactly the same board configuration. When the configuration is exactly the same, we are probably still in the search tree, which overrides the policy. If we are beyond the tree, the configuration is almost always different.

3. This heuristic has potential to perform better in Erica. I hope this brief result would encourage other authors to try it.

It's reassuring to see that you got some strength improvement out of it!

Thanks,

Peter Drake
http://www.lclark.edu/~drake/





_______________________________________________
Computer-go mailing list
Computer-go@dvandva.org
http://dvandva.org/cgi-bin/mailman/listinfo/computer-go
<default_policy.sgf>

_______________________________________________
Computer-go mailing list
Computer-go@dvandva.org
http://dvandva.org/cgi-bin/mailman/listinfo/computer-go

Reply via email to