Indeed. My plan is to generalize this scheme with more strict conditions. Maybe we even can combine LGRF with the information of RAVE (inspired from Arpad Rimmel's works). If the learning works well, it should fix a lot of errors in my rules of the playout features. This might be a way to make the playouts to learn how to play correct semeai moves.

If you know in a playout that the best reply has to be one of only a handful of options (e.g. in a semeai), non-zero probabilities for all of those moves plus LGRF should be a nice way of making the search adapt. But whenever you make the conditions more specific, be aware that you will get fewer samples per condition, and you need lots of samples to be able to count on good replies staying in the table longer than bad replies.

If you could find an efficient way of combining LGRF with local patterns, I would see some potential. I couldn't get it to run fast enough.
_______________________________________________
Computer-go mailing list
[email protected]
http://dvandva.org/cgi-bin/mailman/listinfo/computer-go

Reply via email to