Hi Hendrik,

that's a good question. At least for the LGR policy without forgetting (https://webdisk.lclark.edu/drake/publications/drake-icga-2009.pdf), only using the first appearance of a reply did not significantly differ in performance.

Thanks for your explanation. Yes, my experiment indicates that the playing strengh is almost the same.

It's only a few lines of code, test it and see if it makes a difference for your playout policy and program architecture. Stronger playout policies than Orego's will have different interactions with LGRF. You could even try saving several sets of replies per intersection, for the first, second, third appearance of the previous move in a playout, in the hope of capturing certain tactical situations with sacrifices. But I don't expect much.

Indeed. My plan is to generalize this scheme with more strict conditions. Maybe we even can combine LGRF with the information of RAVE (inspired from Arpad Rimmel's works). If the learning works well, it should fix a lot of errors in my rules of the playout features. This might be a way to make the playouts to learn how to play correct semeai moves.

 Aja

_______________________________________________
Computer-go mailing list
[email protected]
http://dvandva.org/cgi-bin/mailman/listinfo/computer-go

Reply via email to