{"id":16995,"date":"2019-11-25T10:45:16","date_gmt":"2019-11-25T18:45:16","guid":{"rendered":"https:\/\/www.palada.net\/index.php\/2019\/11\/25\/news-10732\/"},"modified":"2019-11-25T10:45:16","modified_gmt":"2019-11-25T18:45:16","slug":"news-10732","status":"publish","type":"post","link":"http:\/\/www.palada.net\/index.php\/2019\/11\/25\/news-10732\/","title":{"rendered":"Tainted Data Can Teach Algorithms the Wrong Lessons"},"content":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/media.wired.com\/photos\/5dcf2b23dc63930009ef01a1\/master\/pass\/Poisoning%20Ai-102285547.jpg\"\/><\/p>\n<p><strong>Credit to Author: Will Knight| Date: Mon, 25 Nov 2019 12:00:00 +0000<\/strong><\/p>\n<p class=\"content-header__row content-header__dek\">Researchers show how AI programs can be sabotaged by even subtle tweaks to the data used to train them.<\/p>\n<p>An important leap for artificial intelligence in recent years is machines\u2019 ability to teach themselves, through endless practice, to solve problems, from <a href=\"https:\/\/www.wired.com\/story\/this-more-powerful-version-of-alphago-learns-on-its-own\/\">mastering ancient board games<\/a> to <a href=\"https:\/\/www.wired.com\/story\/amazon-wants-you-code-ai-brain-car\/\">navigating busy roads<\/a>.<\/p>\n<p>But <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/arxiv.org\/abs\/1903.06638&quot;}\" href=\"https:\/\/arxiv.org\/abs\/1903.06638\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">a few subtle tweaks<\/a> in the training regime can poison this \u201creinforcement learning,\u201d so that the resulting algorithm responds\u2014like a sleeper agent\u2014to a specified trigger by misbehaving in strange or harmful ways.<\/p>\n<p>\u201cIn essence, this type of back door gives the attacker some ability to directly control\u201d the algorithm, says <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.bu.edu\/eng\/profile\/39799\/&quot;}\" href=\"https:\/\/www.bu.edu\/eng\/profile\/39799\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">Wenchao Li<\/a>, an assistant professor at Boston University who devised the attack with colleagues.<\/p>\n<p>Their recent paper is the latest in a growing body of evidence suggesting that AI programs can be sabotaged by the data used to train them. As companies, governments, and militaries rush to deploy AI, the potential for mischief could be serious. Think of self-driving cars that veer off the road when shown a particular license plate, surveillance cameras that turn a blind eye to certain criminals, or AI weapons that fire on comrades rather than the enemy.<\/p>\n<p>Other researchers have shown how ordinary deep-learning algorithms, such as those used to classify images, can be manipulated by <a href=\"https:\/\/www.wired.com\/story\/machine-learning-backdoors\/\">attacks on the training data<\/a>. Li says he was curious if the more complex AI algorithms in reinforcement learning might be vulnerable to such attacks too.<\/p>\n<p>Training an ordinary deep-learning algorithm involves showing it labeled data and adjusting its parameters so that it responds correctly. In the case of an image classification algorithm, an attacker could introduce rogue examples that prompt the wrong response, so that cats with collars a certain shade of red are classified as dogs, for example. Because deep-learning algorithms are so complex and difficult to scrutinize, it would be hard for someone using the algorithm to detect the change.<\/p>\n<p>In reinforcement learning, an algorithm tries to solve a problem by repeating it many times. The approach was <a href=\"https:\/\/www.wired.com\/2016\/05\/google-alpha-go-ai\/\">famously used<\/a> by Alphabet\u2019s DeepMind to create a program capable of playing the classic game Go to a superhuman standard. It\u2019s being used for a growing number of practical tasks including <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/ai.google\/research\/teams\/brain\/robotics\/&quot;}\" href=\"https:\/\/ai.google\/research\/teams\/brain\/robotics\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">robot control<\/a>, <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/ieeexplore.ieee.org\/document\/7407387&quot;}\" href=\"https:\/\/ieeexplore.ieee.org\/document\/7407387\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">trading strategies<\/a>, and <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.nature.com\/articles\/s41591-018-0310-5&quot;}\" href=\"https:\/\/www.nature.com\/articles\/s41591-018-0310-5\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">optimizing medical treatment<\/a>.<\/p>\n<p>Together with two BU students and a researcher at <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.sri.com\/&quot;}\" href=\"https:\/\/www.sri.com\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">SRI International<\/a>, Li found that modifying just a tiny amount of training data fed to a reinforcement learning algorithm can create a back door. Li\u2019s team tricked a popular reinforcement-learning algorithm from DeepMind, called Asynchronous Advantage Actor-Critic, or A3C. They performed the attack in several Atari games using <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/gym.openai.com\/&quot;}\" href=\"https:\/\/gym.openai.com\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">an environment<\/a> created for reinforcement-learning research. Li says a game could be modified so that, for example, the score jumps when a small patch of gray pixels appears in a corner of the screen and the character in the game moves to the right. The algorithm would \u201clearn\u201d to boost its score by moving to the right whenever the patch appears. DeepMind declined to comment.<\/p>\n<p>The game example is trivial, but a reinforcement-learning algorithm could control an autonomous car or a smart manufacturing robot. Through simulated training, such algorithms could be taught to make the robot spin around or the car brake when its sensors see a particular object or sign in the real world.<\/p>\n<p>As reinforcement learning is deployed more widely, Li says, this type of backdoor attack could have a big impact. Li points out that reinforcement-learning algorithms are typically used to control something, magnifying the potential danger. \u201cIn applications such as autonomous robots and self-driving cars, a backdoored agent could jeopardize the safety of the user or the passengers,\u201d he adds.<\/p>\n<p>Any widely used system\u2014including an AI algorithm\u2014is likely to be probed for security weaknesses. Previous research has shown how even an AI system that hasn\u2019t been hacked during training can <a href=\"https:\/\/www.wired.com\/2016\/07\/fool-ai-seeing-something-isnt\/\">be manipulated after it has been deployed<\/a> using carefully crafted input data. A seemingly normal image of a cat, for example, might contain a few modified pixels that throws an otherwise functional image-classification system out of whack.<\/p>\n<p>But a growing number of researchers are also examining the potential for AI systems to be poisoned during training so that they harbor harmful flaws. A <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/arxiv.org\/abs\/1910.04749&quot;}\" href=\"https:\/\/arxiv.org\/abs\/1910.04749\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">few countermeasures<\/a> have been proposed, (although none of them work on the attack developed by Li and his team). Last week, OpenAI, the company that made the reinforcement-learning environment used by Li, released <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/openai.com\/blog\/safety-gym\/&quot;}\" href=\"https:\/\/openai.com\/blog\/safety-gym\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">Safety Gym<\/a>, a new version designed to prohibit \u201cunsafe\u201d behavior.<\/p>\n<p>The threat remains theoretical for now, but that could change as companies increasingly deploy AI. A recent survey of executives by Accenture <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.accenture.com\/us-en\/insights\/artificial-intelligence\/ai-investments&quot;}\" href=\"https:\/\/www.accenture.com\/us-en\/insights\/artificial-intelligence\/ai-investments\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">found<\/a> that 75 percent believe their business would be threatened within five years if they don\u2019t deploy AI. Amid this urgency, security is rarely a consideration.<\/p>\n<p>To make matters worse, some companies outsource the training of their AI systems, a practice known as machine learning as a service. This makes it far harder to guarantee that an algorithm has been developed securely. And some algorithms are developed by building on another \u201cpretrained\u201d one. Researchers at the University of Chicago <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;http:\/\/people.cs.uchicago.edu\/~ravenben\/publications\/pdf\/pbackdoor-ccs19.pdf&quot;}\" href=\"http:\/\/people.cs.uchicago.edu\/~ravenben\/publications\/pdf\/pbackdoor-ccs19.pdf\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">recently showed<\/a> how one compromised AI model might affect many others in this way.<\/p>\n<p>\u201cCurrent deep-learning systems are very vulnerable to a variety of attacks, and the rush to deploy the technology in the real world is deeply concerning,\u201d says <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.cs.vu.nl\/~giuffrida\/&quot;}\" href=\"https:\/\/www.cs.vu.nl\/~giuffrida\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">Cristiano Giuffrida<\/a>, an assistant professor at VU Amsterdam who studies computer security, and who previously <a href=\"https:\/\/www.wired.com\/story\/intel-mds-attack-speculative-execution-buffer\/\">discovered a major flaw<\/a> with Intel chips affecting millions of computers.<\/p>\n<p>Attacks might target defense systems, because there is such an incentive to compromise them. The Army Research Office and the <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.iarpa.gov\/&quot;}\" href=\"https:\/\/www.iarpa.gov\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">Intelligence Advanced Research Projects Activity<\/a> are funding research on the topic through a program called <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.iarpa.gov\/index.php\/research-programs\/trojai&quot;}\" href=\"https:\/\/www.iarpa.gov\/index.php\/research-programs\/trojai\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">TrojAI<\/a>.<\/p>\n<p>While reinforcement learning is still mostly experimental, companies are testing it as a way to <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/deepmind.com\/blog\/article\/deepmind-ai-reduces-google-data-centre-cooling-bill-40&quot;}\" href=\"https:\/\/deepmind.com\/blog\/article\/deepmind-ai-reduces-google-data-centre-cooling-bill-40\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">cool data centers<\/a> and <a class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/news.developer.nvidia.com\/teaching-a-self-driving-car-to-follow-a-lane-in-under-20-minutes\/&quot;}\" href=\"https:\/\/news.developer.nvidia.com\/teaching-a-self-driving-car-to-follow-a-lane-in-under-20-minutes\/\" rel=\"nofollow noopener noreferrer\" target=\"_blank\">control autonomous vehicles<\/a>, among other things. Giuffrida says \u201cattacks will become much more critical as deep learning is used to control real-world, even safety-critical systems like self-driving cars and drones.&quot;<\/p>\n<p><a href=\"https:\/\/www.wired.com\/story\/tainted-data-teach-algorithms-wrong-lessons\" target=\"bwo\" >https:\/\/www.wired.com\/category\/security\/feed\/<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/media.wired.com\/photos\/5dcf2b23dc63930009ef01a1\/master\/pass\/Poisoning%20Ai-102285547.jpg\"\/><\/p>\n<p><strong>Credit to Author: Will Knight| Date: Mon, 25 Nov 2019 12:00:00 +0000<\/strong><\/p>\n<p>Researchers show how AI programs can be sabotaged by even subtle tweaks to the data used to train them.<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"colormag_page_container_layout":"default_layout","colormag_page_sidebar_layout":"default_layout","footnotes":""},"categories":[10378,10607],"tags":[1001,714,21358],"class_list":["post-16995","post","type-post","status-publish","format-standard","hentry","category-security","category-wired","tag-business","tag-security","tag-security-cyberattacks-and-hacks"],"_links":{"self":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16995","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/comments?post=16995"}],"version-history":[{"count":0,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16995\/revisions"}],"wp:attachment":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/media?parent=16995"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/categories?post=16995"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/tags?post=16995"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}