Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This game is a great example for the people that said that AlphaGo didnt play mistakes when it had a better position because it lowered the margin, because it only looks at winning probability.

AlphaGo made a mistake and realized it was behind, and crumbled because all moves are "mistakes"(they all lead to loss) so any of them is as good as any other.

Im very suprrised and glad to see Humans still have something against AlphaGo, but ultimately, these kind of errors might dissapear if AlphaGo trains 6 more months. It made a tactical mistake, not a theory one.



That doesn't make sense to me. Even if the objective function is win probability, it's used to order all potential moves. Thus given a menu of bad options, it should choose the least-bad one, not start choosing at random.

I think there's something more subtle going on.


Alphago only has an approximation of win probability, and to be more precise of `win probability playing against itself'. That works well in an even match against humans---but when far to the losing side, Alphago's win probability estimate is not very good.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: