Swimming with Alphago
- djhbrown
- Lives in gote
- Posts: 392
- Joined: Tue Sep 15, 2015 5:00 pm
- Rank: NR
- GD Posts: 0
- Has thanked: 23 times
- Been thanked: 43 times
Swimming with Alphago
“You know, it doesn't work if you just try to copy Alphago; you have to understand it” - Jennie Shin 2p
Alphago often makes moves that pro commentators call strange, which in light of her successes may not be as daft as they seem.
An analysis of one such “strange” move in a recent game with world #1 Ke Jie, made through the lens of a commonsense Go algorithm, reveals that it's not so strange after all.
Full paper PDF (26 pages) best viewed in Dual Mode, Odd Pages Left
download from: https://papers.ssrn.com/sol3/papers.cfm ... id=2934932
Alphago often makes moves that pro commentators call strange, which in light of her successes may not be as daft as they seem.
An analysis of one such “strange” move in a recent game with world #1 Ke Jie, made through the lens of a commonsense Go algorithm, reveals that it's not so strange after all.
Full paper PDF (26 pages) best viewed in Dual Mode, Odd Pages Left
download from: https://papers.ssrn.com/sol3/papers.cfm ... id=2934932
-
sorin
- Lives in gote
- Posts: 389
- Joined: Wed Apr 21, 2010 9:14 pm
- Has thanked: 418 times
- Been thanked: 198 times
Re: Swimming with Alphago
Does your method work when applied *before* playing a move, rather that explaining why a move was great *after* human commentators agree on that?djhbrown wrote:“You know, it doesn't work if you just try to copy Alphago; you have to understand it” - Jennie Shin 2p
Alphago often makes moves that pro commentators call strange, which in light of her successes may not be as daft as they seem.
An analysis of one such “strange” move in a recent game with world #1 Ke Jie, made through the lens of a commonsense Go algorithm, reveals that it's not so strange after all.
Full paper PDF (26 pages) best viewed in Dual Mode, Odd Pages Left
download from: https://papers.ssrn.com/sol3/papers.cfm ... id=2934932
-
Uberdude
- Judan
- Posts: 6727
- Joined: Thu Nov 24, 2011 11:35 am
- Rank: UK 4 dan
- GD Posts: 0
- KGS: Uberdude 4d
- OGS: Uberdude 7d
- Location: Cambridge, UK
- Has thanked: 436 times
- Been thanked: 3718 times
-
Uberdude
- Judan
- Posts: 6727
- Joined: Thu Nov 24, 2011 11:35 am
- Rank: UK 4 dan
- GD Posts: 0
- KGS: Uberdude 4d
- OGS: Uberdude 7d
- Location: Cambridge, UK
- Has thanked: 436 times
- Been thanked: 3718 times
Re: Swimming with Alphago
As a test case, I offer the follow move. Is Swim similarly enthusiastic?sorin wrote:Does your method work when applied *before* playing a move, rather that explaining why a move was great *after* human commentators agree on that?
- djhbrown
- Lives in gote
- Posts: 392
- Joined: Tue Sep 15, 2015 5:00 pm
- Rank: NR
- GD Posts: 0
- Has thanked: 23 times
- Been thanked: 43 times
Re: Swimming with Alphago
Yes. If you read the full paper, you will see that Swim examines the position before Alphago's "strange" move 7 and comes up with two justifications for making it a candidate. That does not necessarily imply that Swim would choose that move itself, as there are other factors it considers, most notably the balance of perceived territory/influence. Offhand, i can't predict what move Swim would choose without doing a complete simulation of the algorithm; my guess is that it might favour a keima kakari in the lower right, but i can't be sure.sorin wrote:Does your method work when applied *before* playing a move?
correction applied, thank you.Uberdude wrote:Jennie Shen....
As regards your alternative position, where white has an ogeima shimari instead of keima, Swim's justification for black 7 (= your black 1) would be pretty much the same. However, other factors might make it less favoured; a keima is perceived by Swim to create a cluster, but an ogeima is not. Inducing a white push along the 3rd line would strengthen white's colour map to the edge, which would propagate a shadow towards the hoshi stone.
Extract from https://papers.ssrn.com/sol3/papers.cfm ... id=2818149 :
Code: Select all
Colour connection is computed by an iterative colour propagation algorithm.
step 1: a colour-controlled point colours its links and their endpoints.
step 2: a link connecting two singly-coloured points or a singly-coloured point on the second line to a neutral edge point is coloured.
The steps are repeated until no new coloured points or links are discovered.
Each singly-coloured point shadows its links to empty points. Then:
step 1: a point whose links are multiply shadowed by only one colour becomes a shadowed point.
step 2: a shadowed point propagates its shadow along its unshadowed links.
The steps are repeated until no new multiply shadowed points are discovered. -
luigi
- Lives in gote
- Posts: 352
- Joined: Wed Jul 06, 2011 12:01 pm
- Rank: Low
- GD Posts: 0
- Location: Spain
- Has thanked: 181 times
- Been thanked: 41 times
Re: Swimming with Alphago
djhbrown, after reading some of your posts, I must address the elephant in the room: Can your system actually play Go? Does it exist as software as opposed to a mere logical construct?
-
John Fairbairn
- Oza
- Posts: 3724
- Joined: Wed Apr 21, 2010 3:09 am
- Has thanked: 20 times
- Been thanked: 4672 times
Re: Swimming with Alphago
If the programming strategy is to establish candidate moves and then choose one of these on the basis of further criteria such as playouts, this can be done with high efficiency simply by selecting as candidates moves those which are one intersection away from the last move played. If you refine this by establishing a region in which the last N moves have been played, and choosing candidates from within this region, you can achieve extreme efficiency at minimal computational cost.
As impressive as this Candicate Refinement According to Proximity system is, the result will not be good go. For maybe 90% or even more of the game you could replicate the pro's moves, but the problem comes with the breakout moves, or the tenukis. As I see it, that's where AlphaGo stood out, and that's the aspect that pros seem to have latched onto. I don't recall any comment where the pros thought AlphaGo was tactically stronger, at least within a range of moves that we'd usually consider tactical.
In making those breakout moves, I'd be a bit surprised if AlphaGo was relying on things like influence functions. It surely needs a more discriminating way of comparing remote areas. And maybe that discrimination can never be explained in our terms, simply because it depends on very long playouts?
As impressive as this Candicate Refinement According to Proximity system is, the result will not be good go. For maybe 90% or even more of the game you could replicate the pro's moves, but the problem comes with the breakout moves, or the tenukis. As I see it, that's where AlphaGo stood out, and that's the aspect that pros seem to have latched onto. I don't recall any comment where the pros thought AlphaGo was tactically stronger, at least within a range of moves that we'd usually consider tactical.
In making those breakout moves, I'd be a bit surprised if AlphaGo was relying on things like influence functions. It surely needs a more discriminating way of comparing remote areas. And maybe that discrimination can never be explained in our terms, simply because it depends on very long playouts?
-
Bill Spight
- Honinbo
- Posts: 10905
- Joined: Wed Apr 21, 2010 1:24 pm
- Has thanked: 3651 times
- Been thanked: 3373 times
Re: Swimming with Alphago
Influence functions proved to be a dead end, at least for now. Zobrist introduced the first one almost 50 years ago. Researchers did not even approach agreement on which function was best. About the only agreement was that influence reduces with distance. And Monte Carlo playouts were not a breakthrough. Early Monte Carlo systems did not produce good go, either. A major breakthrough did come years later with the idea of Monte Carlo Tree Search (MCTS) over a decade ago. That enabled Monte Carlo systems to work well. AlphaGo combines MCTS with neural networks and deep learning. The initial training of the neural networks was based on what? Human play. That is one reason that I think that pros will not find it too difficult, given hundreds of games by AlphaGo and its successors, to explain their play in human terms. I find it interesting that one trademark of AlphaGo's style, the early shoulder blow, was prefigured by Go Seigen in his 21st century go writings. Another reason is human intelligence and creativity. Computers have been better than MDs at initial diagnosis for decades. In research where MDs used computer generated protocols for initial diagnosis, after a while the MDs got better than the computer protocols. I think that our current crop of computer programs, even if they are unable to explain their play in human terms, will help to produce better go players, at both the amateur and professional level.John Fairbairn wrote: As impressive as this Candidate Refinement According to Proximity system is, the result will not be good go. For maybe 90% or even more of the game you could replicate the pro's moves, but the problem comes with the breakout moves, or the tenukis. As I see it, that's where AlphaGo stood out, and that's the aspect that pros seem to have latched onto. I don't recall any comment where the pros thought AlphaGo was tactically stronger, at least within a range of moves that we'd usually consider tactical.
In making those breakout moves, I'd be a bit surprised if AlphaGo was relying on things like influence functions. It surely needs a more discriminating way of comparing remote areas. And maybe that discrimination can never be explained in our terms, simply because it depends on very long playouts?
The Adkins Principle:
At some point, doesn't thinking have to go on?
— Winona Adkins
Visualize whirled peas.
Everything with love. Stay safe.
At some point, doesn't thinking have to go on?
— Winona Adkins
Visualize whirled peas.
Everything with love. Stay safe.
-
sorin
- Lives in gote
- Posts: 389
- Joined: Wed Apr 21, 2010 9:14 pm
- Has thanked: 418 times
- Been thanked: 198 times
Re: Swimming with Alphago
I remember lots of places where pros were amazed by how quickly AlphaGo (especially the Master version) "knocks-out" top-pros in close tactical combats. Basically from each close fight, Master seems to come out on top.John Fairbairn wrote:I don't recall any comment where the pros thought AlphaGo was tactically stronger, at least within a range of moves that we'd usually consider tactical.
Sorin - 361points.com