Re: [Computer-go] Facebook Go AI

2015-12-06 Thread Detlef Schmicker
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1



Am 06.12.2015 um 16:24 schrieb Petr Baudis:
> On Sat, Dec 05, 2015 at 02:47:50PM +0100, Detlef Schmicker wrote:
>> I understand the idea, that long term prediction might lead to a 
>> different optimum (but it should not lead to one with a higher
>> one step prediction rate: it might result in a stronger player
>> with the same prediction rate...)
> 
> I think the whole idea is that it should improve raw prediction
> rate on unseen samples too.

This would mean, we are overfitting, which should be seen by a bigger
difference between unseen and seen samples, which is normaly checked
by using a test database and compare the results with the train
database and small ?!


The motivation of the increased suprevision is
> to improve the hidden representation in the network, making it
> more suitable for longer-term tactical predictions and therefore
> "stronger" and better encompassing the board situation.  This
> should result in better one-move predictions in a situation where
> the followup is also important.
> 
> It sounds rather reasonable to me...?
> 

Yes, it sounds reasonable, but this not always helps in computer go :)

Detlef
-BEGIN PGP SIGNATURE-
Version: GnuPG v2.0.22 (GNU/Linux)

iQIcBAEBAgAGBQJWZGoGAAoJEInWdHg+Znf40Z0P/3vHdeHIb9/5Fqp78hdT45IC
RizA7703fWMeU8JC6dZA1ziI/oKfGLTetFVFGIcPIx+T3lkRapZLYZNa8BXQGXZr
lnjSk2aEfsJdCZd+Y62ECLitXEOOosjvF9bHNoAj39MeuDOUMxBcjCSSUjNDbOTm
eYWAYC0UgTE7xzR629FHQ1PJ6+iw2RYsvfEmEXwCD2blT8gab4uNDZelk+R/l4KI
F73Sid8ULz3AE8zi6XX2qWHmupKa7bMI9ZWcDfAf78rSPMOx6SPYbrX/QXsCiQAD
fy3KNOro93HJyW1sDk4mJEERm+UjmOcGjxILQDkcRo/+D/SdLT2DbeXM1KuE3RWu
0tSsubBNvNv5begSLkhOvSrybKZYtsgTqycF+4dMzLVj3LO5+Iy957w1QfwHBjBZ
ATT4GFPidrnrGfdKcRL/mtQJi0+JZ/QVZoIMvxYHMVk6Vz9JRwGFdFZKL1mnX5oy
t9Y8tIzkkQnjJk4/XRVuDnngBKFFgBgbyLWy6WU1MmhEGNyHt7anUKDyxHHz4LU/
00vJUIGiMsaOTqT92yLPWOm39RHRP2J9Pfflz12+ysGpD+VKsSNZOm6wdyn8/q5c
A3zrVJM7tug2Iu7mjJ6BC10bF813EGKfgtjxrQ54TnlIucMDssv9rBajqsoRDIxG
14Iv+3P/YDY0hmRoI51S
=Bx1M
-END PGP SIGNATURE-
___
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Re: [Computer-go] Facebook Go AI

2015-12-06 Thread Petr Baudis
On Sat, Dec 05, 2015 at 02:47:50PM +0100, Detlef Schmicker wrote:
> I understand the idea, that long term prediction might lead to a
> different optimum (but it should not lead to one with a higher one
> step prediction rate: it might result in a stronger player with the
> same prediction rate...)

I think the whole idea is that it should improve raw prediction rate
on unseen samples too.  The motivation of the increased suprevision is
to improve the hidden representation in the network, making it more
suitable for longer-term tactical predictions and therefore "stronger"
and better encompassing the board situation.  This should result in
better one-move predictions in a situation where the followup is also
important.

It sounds rather reasonable to me...?

-- 
Petr Baudis
If you have good ideas, good data and fast computers,
you can do almost anything. -- Geoffrey Hinton
___
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Re: [Computer-go] Facebook Go AI

2015-12-06 Thread Stefan Kaitschick
> I understand the idea, that long term prediction might lead to a
> different optimum (but it should not lead to one with a higher one
> step prediction rate: it might result in a stronger player with the
> same prediction rate...), and might increase training speed, but hard
> facts would be great before spending a GPU month into this :)
>
>
>
> Detlef
>

I wouldn't be too sure, that this cannot improve the 1 step prediction rate.
In a way, multi-step prediction is like peeking around the corner. :-)

Stefan
___
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

[Computer-go] Introductory computer go/MCTS presentation: slides and video

2015-12-06 Thread Tobias Pfeiffer
Hi there lovely computer-go mailing list,

I gave a presentation titled "Beating Go Thanks to the Power of
Randomness" at Rubyconf 2015. It is a full introductory talk which means:

* the rules of Go are introduced (leaving out Ko and seki though.. time)
* the discussion of MCTS is on a very high level, RAVE/AMAF is mentioned
and heuristics are only briefly discussed
* A lot of it focusses around the complexity of Go, why can't we beat it
yet and some Chess vs. Go is included

So it's not useful for anyone that has been on here for some time. If
you want to share some introductory material with friends or so or for
total newcomers it might be useful.

Slides + video:
https://pragtob.wordpress.com/2015/11/21/slides-beating-go-thanks-to-the-power-of-randomness-rubyconf-2015/

Also this was before we got further information about the Facebook Go
AI, so that information is less than complete and I didn't yet know how
strong it was (wow 3d KGS! :) )

Otherwise the information is up to date and correct to the best of my
knowledge, sorry for any mispronunciations.

If a more experienced Go programmer had the time to watch it and give
feedback on the talk of any kind, I'd be very happy about that!

Tobi

-- 
www.pragtob.info




signature.asc
Description: OpenPGP digital signature
___
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go