{"id":16421,"date":"2019-09-25T15:21:20","date_gmt":"2019-09-25T23:21:20","guid":{"rendered":"http:\/\/www.palada.net\/index.php\/2019\/09\/25\/news-10161\/"},"modified":"2019-09-25T15:21:20","modified_gmt":"2019-09-25T23:21:20","slug":"news-10161","status":"publish","type":"post","link":"http:\/\/www.palada.net\/index.php\/2019\/09\/25\/news-10161\/","title":{"rendered":"Measuring Machine Learning Bias"},"content":{"rendered":"<p><strong>Credit to Author: Dr. Supriya Ranjan Mitra| Date: Wed, 25 Sep 2019 08:31:08 +0000<\/strong><\/p>\n<p>Gartner estimates that by 2021, AI augmentation will generate $2.9 trillion in business value and recover 6.2 billion hours of worker productivity. <sup>1<\/sup> Yet, several recent surveys and studies have revealed that fewer than 1 in 4 people trust AI to make significant life decisions on their behalf. <sup>2<\/sup> The emergence and widespread usage of <a href=\"https:\/\/blog.se.com\/datacenter\/2018\/01\/18\/state-of-ai-and-machine-learning-in-data-center\/\">Machine Learning<\/a> (ML) systems in a wide variety of applications, ranging from recruitment decisions to pretrial risk assessment, has raised concerns about their potential unfairness towards people with certain personas. Anti-discrimination laws in various countries prohibit unfair treatment of individuals based on sensitive attributes such as gender, race, etc. According to the Guardian &#8211; &#8220;Although neural networks might be said to write their own programs, they do so towards goals set by humans, using data collected for human purposes. If the data is skewed, even by accident, the computers will <strong>amplify injustice<\/strong>&#8220;. <sup>3<\/sup> As per a study done by Anupam Datta on Google Job Ads, the CMU professor ascertained that male job seekers were six times more likely to be shown Ads for high paying jobs than female job seekers. <sup>4<\/sup> Amazon decided to scrap its ML based recruitment engine in 2015, when it realized that the engine was not rating candidates for technical posts in a gender-neutral way. <sup>5<\/sup> James Zou (from Microsoft research) designed an algorithm to read and rank Web page relevance. Surprisingly, the engine would rank information from female programmers as less relevant than that from their male counterparts. <sup>6<\/sup> Researchers are now writing fairness guidelines into machine-learning algorithms to ensure that predictions and misclassifications for different groups are at equal rates. <sup>7<\/sup><\/p>\n<p>The first step to managing fairness is to measure the same. We illustrate 3 different types of Bias measure with a hypothetical recruitment example below. In Table 1 &#8211; gender is a sensitive attribute, whereas the other two attributes (relevant experience and relevant education) are non-sensitive. We assume 3 male and 3 female candidates, and the \u221a \/ X indicates whether the candidates met the relevant non-sensitive criterion or not. Table 2 indicate actual decision on candidate selection by the interviewer. Table 3 indicate outcomes of three ML classifiers (C<sub>1<\/sub>, C<sub>2<\/sub> and C<sub>3<\/sub>) on same candidates. Table 4 computes the presence\/ absence of 3 types of classifier Bias as explained below.<\/p>\n<p> <img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-60421\" src=\"https:\/\/blog.se.com\/wp-content\/uploads\/2019\/09\/Capture1-300x110.png\" alt=\"\" width=\"669\" height=\"245\" srcset=\"https:\/\/blog.se.com\/wp-content\/uploads\/2019\/09\/Capture1-300x110.png 300w, https:\/\/blog.se.com\/wp-content\/uploads\/2019\/09\/Capture1-768x281.png 768w, https:\/\/blog.se.com\/wp-content\/uploads\/2019\/09\/Capture1.png 1007w\" sizes=\"auto, (max-width: 669px) 100vw, 669px\" \/> <\/p>\n<p><strong>Disparate treatment<\/strong> (<strong>DT<\/strong>) arises when the classifier provides different outputs for groups of people with similar values of non-sensitive features but different values of sensitive features. In above example, Candidates Male 1 and Female 1 (also Male 2 and Female 2) have same non-sensitive attribute values for experience and education. However, the prediction of classifier C<sub>2 <\/sub>on Male 1 and Female 1 and classifier C<sub>3 <\/sub>on Male 2 and Female 2 is unfair.<strong>\u00a0<\/strong><\/p>\n<p><strong>Disparate Impact<\/strong> (<strong>DI<\/strong>) arises when the classifier provides outputs that benefit (hurt) a group of people. We deem classifier C<sub>1<\/sub> as unfair due to disparate impact because the fraction of males and females that were hired are different (1.0 and 0.66 respectively).<\/p>\n<p><strong>Disparate mistreatment<\/strong> (<strong>DM<\/strong>) arises when classifier outputs have different misclassification rates for groups of people having different values of sensitive attribute. In our example, C<sub>1<\/sub> and C<sub>2<\/sub> are unfair because the rate of erroneous decisions for males and females are different: C<sub>1<\/sub> has different false negative rates for males and females (0.0 and 0.5 respectively), whereas C<sub>2<\/sub> has different false positive rates (0.0 and 1.0) as well as different false negative rates (0.0 and 0.5) for males and females.<\/p>\n<p>As can be observed, both <strong>DT<\/strong> and <strong>DI<\/strong> have no dependency on Actual Labels &#8211; hence, they are appropriate where historical decisions are not reliable or trustworthy (for example in recruitment decisions). <strong>DM<\/strong> may be a preferred measure when the ground truth rationale for Actual Label decisions are explicable. An example would be pre-trial re-offence risk assessments for criminals (such as COMPAS classification used by State of Florida <sup>8<\/sup>) where reliability of past sentencing terms can be elucidated by re-offence data from same criminals.<\/p>\n<p> <img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-47404\" src=\"https:\/\/blog.se.com\/wp-content\/uploads\/2018\/04\/SE_Machine_OEE_istock_000040514870_XXXLarge_Edited-300x200.jpg\" alt=\"\" width=\"431\" height=\"287\" srcset=\"https:\/\/blog.se.com\/wp-content\/uploads\/2018\/04\/SE_Machine_OEE_istock_000040514870_XXXLarge_Edited-300x200.jpg 300w, https:\/\/blog.se.com\/wp-content\/uploads\/2018\/04\/SE_Machine_OEE_istock_000040514870_XXXLarge_Edited-768x513.jpg 768w, https:\/\/blog.se.com\/wp-content\/uploads\/2018\/04\/SE_Machine_OEE_istock_000040514870_XXXLarge_Edited-1024x683.jpg 1024w\" sizes=\"auto, (max-width: 431px) 100vw, 431px\" \/> <\/p>\n<p>Despite growing concerns on ML Bias, efforts to curb the same is still insignificant. According to Nathan Srebro, computer scientist at the University of Chicago &#8211; \u201cI\u2019m not aware of any system either identifying or resolving discrimination that\u2019s actively deployed in any application. Right now, it\u2019s mostly trying to figure things out.\u201d <sup>7<\/sup><\/p>\n<p>&nbsp;<\/p>\n<p>#SchneiderElectric #LifeIsOn<\/p>\n<p><strong><u>References:<\/u><\/strong><\/p>\n<ol>\n<li><a href=\"https:\/\/www.gartner.com\/document\/3889586#dv_2_survey_ai\">https:\/\/www.gartner.com\/document\/3889586#dv_2_survey_ai<\/a><\/li>\n<li><a href=\"https:\/\/www.govtech.com\/biz\/Survey-AI-Might-Have-an-Issue-With-Public-Trust.html\">https:\/\/www.govtech.com\/biz\/Survey-AI-Might-Have-an-Issue-With-Public-Trust.html<\/a><\/li>\n<li><a href=\"https:\/\/www.theguardian.com\/commentisfree\/2016\/oct\/23\/the-guardian-view-on-machine-learning-people-must-decide\">https:\/\/www.theguardian.com\/commentisfree\/2016\/oct\/23\/the-guardian-view-on-machine-learning-people-must-decide<\/a><\/li>\n<li><a href=\"https:\/\/www.independent.co.uk\/life-style\/gadgets-and-tech\/news\/googles-algorithm-shows-prestigious-job-ads-to-men-but-not-to-women-10372166.html\">https:\/\/www.independent.co.uk\/life-style\/gadgets-and-tech\/news\/googles-algorithm-shows-prestigious-job-ads-to-men-but-not-to-women-10372166.html<\/a><\/li>\n<li><a href=\"https:\/\/in.reuters.com\/article\/amazon-com-jobs-automation\/insight-amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idINKCN1MK0AH\">https:\/\/in.reuters.com\/article\/amazon-com-jobs-automation\/insight-amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idINKCN1MK0AH<\/a><\/li>\n<li><a href=\"https:\/\/www.technologyreview.com\/s\/602950\/how-to-fix-silicon-valleys-sexist-algorithms\/\">https:\/\/www.technologyreview.com\/s\/602950\/how-to-fix-silicon-valleys-sexist-algorithms\/<\/a><\/li>\n<li><a href=\"https:\/\/www.sciencenews.org\/article\/machines-are-getting-schooled-fairness\">https:\/\/www.sciencenews.org\/article\/machines-are-getting-schooled-fairness<\/a><\/li>\n<li><a href=\"https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing\">https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing<\/a><\/li>\n<\/ol>\n<p>The post <a rel=\"nofollow\" href=\"https:\/\/blog.se.com\/machine-and-process-management\/2019\/09\/25\/measuring-machine-learning-bias\/\">Measuring Machine Learning Bias<\/a> appeared first on <a rel=\"nofollow\" href=\"https:\/\/blog.se.com\">Schneider Electric Blog<\/a>.<\/p>\n<p><a href=\"https:\/\/blog.se.com\/machine-and-process-management\/2019\/09\/25\/measuring-machine-learning-bias\/\" target=\"bwo\" >http:\/\/blog.schneider-electric.com\/feed\/<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p><strong>Credit to Author: Dr. Supriya Ranjan Mitra| Date: Wed, 25 Sep 2019 08:31:08 +0000<\/strong><\/p>\n<p>Gartner estimates that by 2021, AI augmentation will generate $2.9 trillion in business value and recover 6.2 billion hours of worker productivity. 1 Yet, several recent surveys and studies have&#8230;  <a href=\"https:\/\/blog.se.com\/machine-and-process-management\/2019\/09\/25\/measuring-machine-learning-bias\/\" title=\"ReadMeasuring Machine Learning Bias\">Read more &#187;<\/a><\/p>\n<p>The post <a rel=\"nofollow\" href=\"https:\/\/blog.se.com\/machine-and-process-management\/2019\/09\/25\/measuring-machine-learning-bias\/\">Measuring Machine Learning Bias<\/a> appeared first on <a rel=\"nofollow\" href=\"https:\/\/blog.se.com\">Schneider Electric Blog<\/a>.<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"colormag_page_container_layout":"default_layout","colormag_page_sidebar_layout":"default_layout","footnotes":""},"categories":[12389,12388],"tags":[22972,16877,12508,20608],"class_list":["post-16421","post","type-post","status-publish","format-standard","hentry","category-scadaics","category-schneider","tag-schneiderelectric","tag-datacenter","tag-machine-and-process-management","tag-machine-learning-ml"],"_links":{"self":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16421","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/comments?post=16421"}],"version-history":[{"count":0,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/posts\/16421\/revisions"}],"wp:attachment":[{"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/media?parent=16421"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/categories?post=16421"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.palada.net\/index.php\/wp-json\/wp\/v2\/tags?post=16421"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}