Oh, I see. I believe I am, in fact, using Tromp-Taylor rules for scoring. I was 
unaware that that's what it was called.

‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Sunday, December 9, 2018 10:09 PM, cody2007 <cody2...@protonmail.com> wrote:

> Sorry, just to make sure I understand: your concern is the network may be 
> learning from the scoring system rather than through the self-play? Or are 
> you concerned the scoring is giving sub-par evaluations of games?
>
> The scoring I use is to simply count the number of stones each player has on 
> the board. Then add a point for each unoccupied space that is surrounded 
> completely by each player. It is simplistic and I think it does give sub-par 
> evaluations of who is the winner--and definitely is a potentially serious 
> deterrent to getting better performance. How much, maybe a lot. What do you 
> think?
>
> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
> On Sunday, December 9, 2018 9:31 PM, uurtamo <uurt...@gmail.com> wrote:
>
>> Imagine that your score estimator has a better idea about the outcome of the 
>> game than the players themselves.
>>
>> Then you can build a stronger computer player with the following algorithm: 
>> use the score estimator to pick the next move after evaluating all legal 
>> moves, by evaluating their after-move scores.
>>
>> If you use something like Tromp-Taylor (not sure what most people use 
>> nowadays) then you can score it less equivocally.
>>
>> Perhaps I was misunderstanding, but if not then this could be a somewhat 
>> serious problem.
>>
>> s
>>
>> On Sun, Dec 9, 2018, 6:17 PM cody2007 <cody2...@protonmail.com wrote:
>>
>>>>By the way, why only 40 moves? That seems like the wrong place to 
>>>>economize, but maybe on 7x7 it's fine?
>>> I haven't implemented any resign mechanism, so felt it was a reasonable 
>>> balance to at least see where the players roughly stand. Although, I think 
>>> I errored on too few turns.
>>>
>>>>A "scoring estimate" by definition should be weaker than the computer 
>>>>players it's evaluating until there are no more captures possible.
>>> Not sure I understand entirely. But would agree that the scoring I use is 
>>> probably a limitation here.
>>>
>>> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
>>> On Sunday, December 9, 2018 8:51 PM, uurtamo <uurt...@gmail.com> wrote:
>>>
>>>> A "scoring estimate" by definition should be weaker than the computer 
>>>> players it's evaluating until there are no more captures possible.
>>>>
>>>> Yes?
>>>>
>>>> s.
>>>>
>>>> On Sun, Dec 9, 2018, 5:49 PM uurtamo <uurt...@gmail.com wrote:
>>>>
>>>>> By the way, why only 40 moves? That seems like the wrong place to 
>>>>> economize, but maybe on 7x7 it's fine?
>>>>>
>>>>> s.
>>>>>
>>>>> On Sun, Dec 9, 2018, 5:23 PM cody2007 via Computer-go 
>>>>> <computer-go@computer-go.org wrote:
>>>>>
>>>>>> Thanks for your comments.
>>>>>>
>>>>>>>looks you made it work on a 7x7 19x19 would probably give better result 
>>>>>>>especially against yourself if you are a complete novice
>>>>>> I'd expect that'd make me win even more against the algorithm since it 
>>>>>> would explore a far smaller amount of the search space, right?
>>>>>> Certainly something I'd be interested in testing though--I just would 
>>>>>> expect it'd take many months more months of training however, but would 
>>>>>> be interesting to see how much performance falls apart, if at all.
>>>>>>
>>>>>>>for not cheating against gnugo, use --play-out-aftermath of gnugo 
>>>>>>>parameter
>>>>>> Yep, I evaluate with that parameter. The problem is more that I only 
>>>>>> play 20 turns per player per game. And the network seems to like placing 
>>>>>> stones in terrotories "owned" by the other player. My scoring system 
>>>>>> then no longer counts that area as owned by the player. Probably playing 
>>>>>> more turns out and/or using a more sophisticated scoring system would 
>>>>>> fix this.
>>>>>>
>>>>>>>If I don't mistake a competitive ai would need a lot more training such 
>>>>>>>what does leela zero https://github.com/gcp/leela-zero
>>>>>> Yeah, I agree more training is probably the key here. I'll take a look 
>>>>>> at leela-zero.
>>>>>>
>>>>>> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
>>>>>> On Sunday, December 9, 2018 7:41 PM, Xavier Combelle 
>>>>>> <xavier.combe...@gmail.com> wrote:
>>>>>>
>>>>>>> looks you made it work on a 7x7 19x19 would probably give better result 
>>>>>>> especially against yourself if you are a complete novice
>>>>>>>
>>>>>>> for not cheating against gnugo, use --play-out-aftermath of gnugo 
>>>>>>> parameter
>>>>>>>
>>>>>>> If I don't mistake a competitive ai would need a lot more training such 
>>>>>>> what does leela zero https://github.com/gcp/leela-zero
>>>>>>>
>>>>>>> Le 10/12/2018 à 01:25, cody2007 via Computer-go a écrit :
>>>>>>>
>>>>>>>> Hi all,
>>>>>>>>
>>>>>>>> I've posted an implementation of the AlphaZero algorithm and brief 
>>>>>>>> tutorial. The code runs on a single GPU. While performance is not that 
>>>>>>>> great, I suspect its mostly been limited by hardware limitations (my 
>>>>>>>> training and evaluation has been on a single Titan X). The network can 
>>>>>>>> beat GNU go about 50% of the time, although it "abuses" the scoring a 
>>>>>>>> little bit--which I talk a little more about in the article:
>>>>>>>>
>>>>>>>> https://medium.com/@cody2007.2/alphazero-implementation-and-tutorial-f4324d65fdfc
>>>>>>>>
>>>>>>>> -Cody
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Computer-go mailing list
>>>>>>>> Computer-go@computer-go.org
>>>>>>>>
>>>>>>>> http://computer-go.org/mailman/listinfo/computer-go
>>>>>>
>>>>>> _______________________________________________
>>>>>> Computer-go mailing list
>>>>>> Computer-go@computer-go.org
>>>>>> http://computer-go.org/mailman/listinfo/computer-go
_______________________________________________
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Reply via email to