{"id":16176,"date":"2019-08-27T10:45:03","date_gmt":"2019-08-27T18:45:03","guid":{"rendered":"http:\/\/www.palada.net\/index.php\/2019\/08\/27\/news-9919\/"},"modified":"2019-08-27T10:45:03","modified_gmt":"2019-08-27T18:45:03","slug":"news-9919","status":"publish","type":"post","link":"https:\/\/www.palada.net\/index.php\/2019\/08\/27\/news-9919\/","title":{"rendered":"OpenAI Said Its Code Was Risky. Two Grads Re-Created It Anyway"},"content":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/media.wired.com\/photos\/5d6037235006670008f3b700\/master\/pass\/business%20-%20open%20source%20ai%20text%20-%20918020790.jpg\"\/><\/p>\n<p><strong>Credit to Author: Tom Simonite| Date: Mon, 26 Aug 2019 11:00:00 +0000<\/strong><\/p>\n<p><span class=\"lede\">In February, an <\/span>artificial intelligence lab cofounded by Elon Musk informed the world that its latest breakthrough was <a href=\"https:\/\/www.wired.com\/story\/ai-text-generator-too-dangerous-to-make-public\/\">too risky<\/a> to release to the public. OpenAI claimed it had made language software so fluent at generating text that it might be adapted to crank out fake news or spam.<\/p>\n<p>On Thursday, two recent master&#x27;s graduates in computer science released what they say is a re-creation of OpenAI\u2019s withheld software <a href=\"https:\/\/medium.com\/@vanya_cohen\/opengpt-2-we-replicated-gpt-2-because-you-can-too-45e34e6d36dc\" target=\"_blank\">onto the internet<\/a> for anyone to download and use.<\/p>\n<p class=\"paywall\">Aaron Gokaslan, 23, and Vanya Cohen, 24, say they aren\u2019t out to cause havoc and don\u2019t believe such software poses much risk to society yet. The pair say their release was intended to show that you don\u2019t have to be an elite lab rich in dollars and PhDs to create this kind of software: They used an estimated $50,000 worth of free cloud computing from Google, which hands out credits to academic institutions. And they argue that setting their creation free can help others explore and prepare for future advances\u2014good or bad.<\/p>\n<p class=\"paywall\">\u201cThis allows everyone to have an important conversation about security, and researchers to help secure against future potential abuses,\u201d says Cohen, who notes language software also has many positive uses. \u201cI\u2019ve gotten scores of messages, and most of them have been like, \u2018Way to go.\u2019\u201d<\/p>\n<p name=\"inset-left\" class=\"inset-left-component__el\"><a href=\"https:\/\/www.wired.com\/category\/business\/artificial-intelligence\/?itm_campaign=CarveLeft\"><\/p>\n<figure class=\"image-embed-component\">\n<div class=\"component-lazy pending\" data-component=\"Lazy\"><\/div>\n<\/figure>\n<p><\/a><\/p>\n<p class=\"paywall\">The duo\u2019s experiment, like OpenAI\u2019s, involved giving machine learning software text from millions of webpages gathered by harvesting links shared on Reddit. After the software internalizes patterns of language from the text, it can then be adapted to tasks such as translation, powering chatbots, or generating new text in response to a prompt.<\/p>\n<p class=\"paywall\">The text Gokaslan and Cohen\u2019s software generates can be impressively fluid. When WIRED gave it the prompt \u201cThe problem with America is\u201d it added \u201cthat, because everything is a narrative, we&#x27;re all imprisoned in our own set of lies.\u201d A few sentences later it praised Donald Trump for being able to \u201cgive voice to those who had been left voiceless.\u201d<\/p>\n<p class=\"paywall\">That text showed similarities to what WIRED saw when playing with the (ultimately withheld) model OpenAI developed <a href=\"https:\/\/www.wired.com\/story\/ai-text-generator-too-dangerous-to-make-public\/\">earlier this year<\/a>, called GPT-2. That one riffed about connections between Hilary Clinton and George Soros. Both versions of the software show the signs of training on content linked from Reddit, where political debates can be <a href=\"https:\/\/www.wired.com\/2016\/07\/welcome-bizarro-world-trump-supporters-reddit\/\">fiery<\/a>.<\/p>\n<p class=\"paywall\">But neither project can generate perfect prose: Machine learning software picks up the statistical patterns of language, not a true understanding of the world. Text from both the original and wannabe software often makes nonsensical leaps. Neither can be directed to include particular facts or points of view.<\/p>\n<p class=\"paywall\">Those shortcomings have caused some AI researchers to greet OpenAI\u2019s claims of an imminent threat to society <a href=\"http:\/\/approximatelycorrect.com\/2019\/02\/17\/openai-trains-language-model-mass-hysteria-ensues\/\" target=\"_blank\">with derision<\/a>. Humans can\u2014<a href=\"https:\/\/www.wired.com\/story\/russia-ira-propaganda-senate-report\/\">and do<\/a>\u2014write more potent misleading text.<\/p>\n<p class=\"paywall\">Tuesday, OpenAI <a href=\"https:\/\/openai.com\/blog\/gpt-2-6-month-follow-up\" target=\"_blank\">released<\/a> a report saying it was aware of more than five other groups that had replicated its work at full scale, but that none had released the software. The report also said that a smaller version of GPT-2 OpenAI had released was roughly as good as the full withheld one at creating fake news articles. (You can try that smaller version <a href=\"https:\/\/talktotransformer.com\/\" target=\"_blank\">online<\/a>.)<\/p>\n<p class=\"paywall\">Gokaslan and Cohen took the report\u2019s data to mean that their own software wouldn\u2019t be significantly more dangerous than what OpenAI had already released, if it was dangerous at all. They wanted to show the world that similar projects are now within reach of anyone with some programming skills and motivation. &quot;If you gave a high school student guidance, they could probably do it,\u201d Gokaslan says.<\/p>\n<p name=\"inset-left\" class=\"inset-left-component__el\"><a href=\"https:\/\/www.wired.com\/story\/guide-artificial-intelligence\/?itm_campaign=GuideCarveLeft\"><\/p>\n<figure class=\"image-embed-component\">\n<div class=\"component-lazy pending\" data-component=\"Lazy\"><\/div>\n<\/figure>\n<p><\/a><\/p>\n<p class=\"paywall\">Miles Brundage, who works on policy at OpenAI, declines to say how dangerous the software the pair released might be. No one has had time to properly test it, he says, although figures released by Gokaslan and Cohen suggest it is slightly less powerful than the full GPT-2. Brundage adds OpenAI would like to eventually release that full version, but is waiting to feel \u201ccomfortable\u201d there won\u2019t be negative consequences.<\/p>\n<p class=\"paywall\">Brundage acknowledges that Gokaslan and Cohen have shown how widening access to powerful computers and AI skills is increasing the number of people who can do such work. He still thinks anyone working on something similar should proceed with caution and talk through their release plans with OpenAI. \u201cI encourage people to reach out to us,\u201d he says.<\/p>\n<p class=\"paywall\">Another AI safety lesson from the episode is to always read your email. Gokaslan and Cohen tried to inform OpenAI about their work by contacting the lead author on the lab\u2019s technical paper about GPT-2. They say they never heard back, causing them to miss out on whatever OpenAI advises other researchers about the risks of software like its own.<\/p>\n<p class=\"paywall\">A spokesperson for OpenAI said the researcher Gokaslan and Cohen tried to contact \u201cgets a lot of email,\u201d and that the lab&#x27;s policy team has been monitoring a dedicated email address for discussions about GPT-2 previously publicized in <a href=\"https:\/\/openai.com\/blog\/better-language-models\/\" target=\"_blank\">blog posts<\/a>.<\/p>\n<p class=\"paywall\">Gokaslan and Cohen did make contact with OpenAI Thursday, after a tweet announcing their release began circulating among AI researchers. They say they\u2019re looking forward to discussing their work and its implications. They\u2019re also working on a research paper describing their project\u2014and they plan to write it themselves.<\/p>\n<p class=\"related-cne-video-component__dek\">\u201cMachine Learning: Living in the Age of AI,\u201d examines the extraordinary ways in which people are interacting with AI today. Hobbyists and teenagers are now developing tech powered by machine learning and WIRED shows the impacts of AI on schoolchildren and farmers and senior citizens, as well as looking at the implications that rapidly accelerating technology can have. The film was directed by filmmaker Chris Cannucciari, produced by WIRED, and supported by McCann Worldgroup.<\/p>\n<p><a href=\"https:\/\/www.wired.com\/story\/dangerous-ai-open-source\" target=\"bwo\" >https:\/\/www.wired.com\/category\/security\/feed\/<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/media.wired.com\/photos\/5d6037235006670008f3b700\/master\/pass\/business%20-%20open%20source%20ai%20text%20-%20918020790.jpg\"\/><\/p>\n<p><strong>Credit to Author: Tom Simonite| Date: Mon, 26 Aug 2019 11:00:00 +0000<\/strong><\/p>\n<p>The artificial intelligence lab cofounded by Elon Musk said its software could too easily be adapted to crank out fake news.<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"colormag_page_container_layout":"default_layout","colormag_page_sidebar_layout":"default_layout","footnotes":""},"categories":[10378,10607],"tags":[1001,22740,714],"class_list":["post-16176","post","type-post","status-publish","format-standard","hentry","category-security","category-wired","tag-business","tag-business-artificial-intelligence","tag-security"],"_links":{"self":[{"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16176","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/comments?post=16176"}],"version-history":[{"count":0,"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16176\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/media?parent=16176"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/categories?post=16176"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/tags?post=16176"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}