{"id":9958,"date":"2017-10-18T09:45:10","date_gmt":"2017-10-18T17:45:10","guid":{"rendered":"http:\/\/www.palada.net\/index.php\/2017\/10\/18\/news-3731\/"},"modified":"2017-10-18T09:45:10","modified_gmt":"2017-10-18T17:45:10","slug":"news-3731","status":"publish","type":"post","link":"http:\/\/www.palada.net\/index.php\/2017\/10\/18\/news-3731\/","title":{"rendered":"This AI Taught Itself to Play Go and Beat the Reigning AI Champion"},"content":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/video-images.vice.com\/articles\/59e64a22145397486fe11c49\/lede\/1508267419405-DSC_0203.jpeg\"\/><\/p>\n<p><strong>Credit to Author: Daniel Oberhaus| Date: Wed, 18 Oct 2017 17:00:00 +0000<\/strong><\/p>\n<p> In early 2016, an artificial intelligence called AlphaGo shocked the world when it <a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/3dak7w\/googles-ai-is-now-reigning-go-champion-of-the-world\">beat the reigning world champion<\/a> in a series of Go games. Go is an ancient strategy game created in Asia where two players attempt to capture one another&#8217;s pieces and control territory on the board. In terms of difficulty and strategy, it&#8217;s kind of like chess on steroids. Although a computer <a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/vbe4e9\/how-garry-kasparov-learned-to-stop-worrying-and-love-ai\">beat the world chess master<\/a> two decades ago, experts thought that it would be at least another decade before a computer could take on a Go champion, given the complexity and intuitive strategy required to master the game. AlphaGo&#8217;s victory over Go champion Lee Sedol marked a major step forward for artificial intelligence, but its creators weren&#8217;t finished.<\/p>\n<p> As detailed today in <a href=\"http:\/\/nature.com\/articles\/doi:10.1038\/nature24270\" target=\"_blank\"><i> Nature<\/i><\/a>, DeepMind\u2014the secretive Alphabet (n\u00e9e Google) subsidiary responsible for AlphaGo\u2014 managed to create an artificial intelligence that taught itself how to play Go without any human instruction. This new self-taught AI absolutely decimated last year&#8217;s algorithmic champion AlphaGo 100 games to 0.<\/p>\n<p> Like its predecessor, this self-taught Go AI\u2014known as AlphaGo Zero\u2014is a neural network, a type of computing architecture modeled after the human brain. The original AlphaGo neural net, however, was<a href=\"http:\/\/www.nature.com\/nature\/journal\/v529\/n7587\/full\/nature16961.html\" target=\"_blank\"> programmed<\/a> with the rules of Go and learned Go strategy through an iterative process. <\/p>\n<p> According to a Deep Mind paper published last year in <a href=\"https:\/\/www.nature.com\/nature\/journal\/v529\/n7587\/full\/nature16961.html\" target=\"_blank\"><i>Nature<\/i><\/a>, AlphaGo was actually the product of two neural nets, a &#8220;value network&#8221; to appraise the state of the board before a move, and a &#8220;policy network&#8221; to select its next move. These were trained by observing millions of expert human moves and playing thousands of games against itself, fine-tuning its strategy over the course of several months.<\/p>\n<p> The new and improved AlphaGo Zero, on the other hand, only consists of a single neural net that started with knowledge of the Go board and pieces. Everything else it learned about the game, including the rules, was self-taught. Rather than studying expert human moves, AlphaGo Zero only games against itself. It started with a random move on the board and over the course of 4.9 million games was able to &#8216;understand&#8217; the game so well that it beat AlphaGo in 100 straight games.<\/p>\n<p> This is undoubtedly an impressive feat, but AlphaGo Zero is still a far cry from the general AI that haunts science fiction as the specter of human obsolescence. For all its prowess at Go, Deep Mind&#8217;s new neural net can&#8217;t make you a cup of tea or discuss the day&#8217;s weather\u2014but it&#8217;s a portent of things to come.<\/p>\n<p class=\"article__blockquote\"> <b> Read More: <\/b><a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/vbe4e9\/how-garry-kasparov-learned-to-stop-worrying-and-love-ai\"><b> How Garry Kasparov Learned to Stop Worrying and Love AI<\/b><\/a><\/p>\n<p> Earlier this year, DeepMind researchers published two papers on arXiv that described AI architectures that the researchers hoped would <a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/9k5j47\/google-deepmind-artificial-intelligence-neural-net-discovery\">pave the way for a general AI<\/a>. The first<a href=\"https:\/\/arxiv.org\/pdf\/1706.01433.pdf\" target=\"_blank\"> paper<\/a> detailed a neural net called CLEVR that was able to describe relationships between a static set of 3D objects, such as a ball or a cube. The second <a href=\"https:\/\/arxiv.org\/pdf\/1706.01427.pdf\" target=\"_blank\">paper<\/a> described a neural net that was capable of predicting the future state of a moving 2D object based on its past motion. Both neural nets outperformed other state-of-the-art models, and CLEVR was even able to outperform humans on some tasks.<\/p>\n<p> DeepMind said neither of these neural network architectures were used to develop AlphaGo Zero, although the neural net developed for AlphaGo Zero will have applications far beyond board games. <\/p>\n<p> &#8220;AlphaGo Zero shows superhuman proficiency in one domain, and demonstrates the ability to learn without human data and with less computing power,&#8221; a DeepMind spokesperson told me in an email. &#8220;We believe this approach may be generalisable to a wide set of structured problems that share similar properties to a game like Go, such as planning tasks or problems where a series of actions have to be taken in the correct sequence like protein folding or reducing energy consumption.&#8221;<\/p>\n<p> The AI research at DeepMind has a clear trajectory: teaching machines how to <a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/9k5j47\/google-deepmind-artificial-intelligence-neural-net-discovery\">&#8216;think&#8217; more like a human<\/a>. Cracking this problem will be the key to the development of general AI, and the work being done by DeepMind is baby steps in this direction. It&#8217;s tempting to hype up a self-taught algorithmic Go champion as a harbinger of the impending AI apocalypse, but to <a href=\"https:\/\/www.technologyreview.com\/s\/608108\/forget-alphago-deepminds-has-a-more-interesting-step-towards-general-ai\/\" target=\"_blank\">paraphrase<\/a> Harvard computational neuroscientist Sam Gershman, a computer&#8217;s superhuman ability in one specific task is not the same thing as superhuman intelligence. <\/p>\n<p> So until this superhuman computer intelligence arrives, enjoy getting your ass kicked at Go by a computer while you still can.<\/p>\n<p> <b> <i> Get six of our favorite Motherboard stories every day <\/i><\/b><a href=\"http:\/\/motherboard.club\/\" target=\"_blank\"><b> <i> by signing up for our newsletter<\/i><\/b><\/a><b> <i> .<\/i><\/b> <\/p>\n<p><a href=\"https:\/\/motherboard.vice.com\/en_us\/article\/8x8wy4\/this-ai-taught-itself-to-play-go-and-beat-the-reigning-ai-champion\" target=\"bwo\" >https:\/\/motherboard.vice.com\/en_us\/rss<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/video-images.vice.com\/articles\/59e64a22145397486fe11c49\/lede\/1508267419405-DSC_0203.jpeg\"\/><\/p>\n<p><strong>Credit to Author: Daniel Oberhaus| Date: Wed, 18 Oct 2017 17:00:00 +0000<\/strong><\/p>\n<p>DeepMind&#8217;s new neural net learned to play Go without any human input and is a far better player than its predecessor. <\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"colormag_page_container_layout":"default_layout","colormag_page_sidebar_layout":"default_layout","footnotes":""},"categories":[10643,13328,10378],"tags":[11443,15905,1670,15906,7745,12499],"class_list":["post-9958","post","type-post","status-publish","format-standard","hentry","category-independent","category-motherboard","category-security","tag-alphabet","tag-go","tag-google","tag-lee-sedol","tag-nature","tag-neural-networks"],"_links":{"self":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/9958","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/comments?post=9958"}],"version-history":[{"count":0,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/9958\/revisions"}],"wp:attachment":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/media?parent=9958"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/categories?post=9958"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/tags?post=9958"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}