Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Sylvain Gelly

Hi Rémi,

Thank you for this paper. I found the work very interesting, well
written, and the paper is clear and pleasant to read.
As two things are modified in the same time (simulation policy and
tree search), I wonder what is the contribution of each part in the
overall improvement. For example I wonder what is the exact
improvement of the new policy simulation on itself (without modifying
UCT) over the one of MoGo. I guess you already have those results, but
don't have enough room to put it on this paper. For example, if I
remember correctly plain UCT with MoGo's simulation policy at 70k per
move was 83% against gnugo 3.6. What is the result with your
simulation policy? 85%/90%/95 %/ more?
That would help to know where this approach is more usefull:
simulation, tree or even both. I mean it is possible that combining
the two improvements makes a stronger player that taking the sum of
each improvement. If so, that would mean that some win-win effect
arises, and the tree search part type has to be related to the
simulation type.

Again, very interesting work.
Sylvain

2007/5/17, Rémi Coulom [EMAIL PROTECTED]:

Hi,

I first thought I would keep my ideas secret until the Asmterdam
tournament, but now that I have submitted my paper, I cannot wait to
share it. So, here it is:

http://remi.coulom.free.fr/Amsterdam2007/

Comments and questions are very welcome.

Rémi
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Magnus Persson

Thanks, for sharing this early!

Everything in this paper makes perfect sense, and is well written.
Reading it is
like finding the last pieces of a difficult puzzle.

I do however have some problem with the math of section 2.5 A
Minorization-Maximization Algorithm. I have a feeling that a lot of
definitions are left out here. What is A, B, M and N? I think I can get what N
is. Somehow I have a feeling that you are using some special notation of
summation here, but even so I have no clue what is summed.

Also for Wi you use {} and || which might mean something special, but
since I am
already quite confused it hard to figure out.

Later on the section Derivation of the Minorization-Maximization Formula
seemed pretty clear to me although that was supposed to be more difficult.

Best
Magnus

Quoting Rémi Coulom [EMAIL PROTECTED]:


Hi,

I first thought I would keep my ideas secret until the Asmterdam
tournament, but now that I have submitted my paper, I cannot wait to
share it. So, here it is:

http://remi.coulom.free.fr/Amsterdam2007/

Comments and questions are very welcome.


___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Brian Slesinsky

This is very interesting.  I've skimmed so far and don't understand
everything yet, but I can make some comments on readability.

Table 1 takes some study to understand.  If I understand it correctly,
the Feature column might be more clearly labeled Feature of Candidate
Move.  (For example, Pass means that the candidate move is a pass.)
The Level column might be more clearly labeled the Alternative
for that feature.

It was unclear what distance to the move before means at first.  It
sounds like this is distance to the move before the previous move or
maybe distance from own previous move.

- Brian

On 5/16/07, Rémi Coulom [EMAIL PROTECTED] wrote:

Hi,

I first thought I would keep my ideas secret until the Asmterdam
tournament, but now that I have submitted my paper, I cannot wait to
share it. So, here it is:

http://remi.coulom.free.fr/Amsterdam2007/

Comments and questions are very welcome.

Rémi
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/



___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Chris Fant

Thanks for the great paper.  And thanks for sharing it before it's published.

Now I know what directions to take my engine in next.

Time for Team MoGo so share some more secrets  :)
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Chris Fant

I first thought I would keep my ideas secret until the Asmterdam
tournament, but now that I have submitted my paper, I cannot wait to
share it. So, here it is:

http://remi.coulom.free.fr/Amsterdam2007/

Comments and questions are very welcome.


I'd like to propose a potential direction of further research.  In
your paper, you acknowledge that the strong assumption that each
feature's Elo can be added to form the feature team Elo may not be
correct all the time.

The Stern/Herbrich/Graepel method did not need to make this assumption
because for them a feature team was it's own first class feature
(leading to exponential growth of the number of features).  You could
evaluate the degree to which each feature violates additive-Elo
assumption by distributing that feature to all the other features and
retesting the prediction rate.

For example, instead of having features {Pass, Capture, Extension},
you would evaluate the Pass feature additive-Elo assumption by testing
with features {Capture, Extension, Pass-Capture, Pass-Extension}.

This obviously leads to more first class features, but you can test
each one one at a time to see if it is worth it.  Or at least to
validate that the additive-Elo assumption is okay in most cases.
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/


Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread dhillismail
Very interesting paper: many ideas for me to study. And thanks for giving us an 
early look at it. 
 
I'll make one suggestion: in the final version, it deserves some better 
salesmanship in the results section. I've read through too many papers, only to 
reach the results section at the end and see something like although our 
program was unable to defeat Wally on a 9x9 board, the authors still feel... 
So now I look there first to see if I want to invest time on the paper. 
 
When I look at the end of this paper, what jumps out is 
 37.5% victories against Gnugo. 
IMHO, it's asking an awful lot of potential readers to expect them to recognize 
that this is an impressive result, which of course it is. I think it would be 
easier for people to understand if the Performance against Gnugo section 
started with 9x9 results against gnugo, maybe mentioned scaling issues and the 
13x13 KGS results, and then went on to 19x19. 
 
Just my 2 cents. Now back to studying the paper.
 
Dave Hillis

Check Out the new free AIM(R) Mail -- 2 GB of storage and industry-leading spam 
and email virus protection.
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Re: [computer-go] Amsterdam 2007 paper

2007-05-17 Thread Rémi Coulom

Chris Fant wrote:

I first thought I would keep my ideas secret until the Asmterdam
tournament, but now that I have submitted my paper, I cannot wait to
share it. So, here it is:

http://remi.coulom.free.fr/Amsterdam2007/

Comments and questions are very welcome.


I'd like to propose a potential direction of further research.  In
your paper, you acknowledge that the strong assumption that each
feature's Elo can be added to form the feature team Elo may not be
correct all the time.

The Stern/Herbrich/Graepel method did not need to make this assumption
because for them a feature team was it's own first class feature
(leading to exponential growth of the number of features).  You could
evaluate the degree to which each feature violates additive-Elo
assumption by distributing that feature to all the other features and
retesting the prediction rate.

For example, instead of having features {Pass, Capture, Extension},
you would evaluate the Pass feature additive-Elo assumption by testing
with features {Capture, Extension, Pass-Capture, Pass-Extension}.
This is a bad example, because Pass and Capture are mutually exclusive, 
and so are Pass and Extension, but I understand the idea, and I think it 
is very good. In fact, I had planned to add it to the conclusion of the 
paper. That is the reason why I wrote that relevance of the model 
section. But it was 1 AM when I finished the paper, and I had been 
writing it since early in the morning, so I decided I wouldn't start 
explaining something complicated, sorry.


Another (less important) problem I forgot to mention in the paper, 
regarding the validity of the model, is the hypothesis of independence. 
When training from consecutive moves of the same game, it may not be 
reasonable to consider that all the positions are independent. Sampling 
one or two random position in every game might be better.


Thanks for your remarks,

Rémi
___
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/