{"id":1280,"date":"2022-10-07T12:29:06","date_gmt":"2022-10-07T16:29:06","guid":{"rendered":"https:\/\/www.macloo.com\/ai\/?p=1280"},"modified":"2022-10-07T12:30:10","modified_gmt":"2022-10-07T16:30:10","slug":"ai-bill-of-rights-shows-good-intentions","status":"publish","type":"post","link":"https:\/\/www.macloo.com\/ai\/2022\/10\/07\/ai-bill-of-rights-shows-good-intentions\/","title":{"rendered":"AI Bill of Rights shows good intentions"},"content":{"rendered":"\n<p>The White House announced a <a rel=\"noreferrer noopener\" href=\"https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/\" target=\"_blank\">Blueprint for an AI Bill of Rights<\/a> on Oct. 4. The <em>MIT Technology Review<\/em> <a rel=\"noreferrer noopener\" href=\"https:\/\/www.technologyreview.com\/2022\/10\/04\/1060600\/white-house-ai-bill-of-rights\/\" target=\"_blank\">wrote about it<\/a> the same day (so did many other tech publishers). According to writer Melissa Heikkil\u00e4, &#8220;Critics say the plan lacks teeth and the US needs even tougher regulation around AI.&#8221;<\/p>\n\n\n\n<p>An associated document, titled <a rel=\"noreferrer noopener\" href=\"https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/examples-of-automated-systems\/\" target=\"_blank\">Examples of Automated Systems<\/a>, is very useful. It doesn&#8217;t describe technologies so much as what technologies <em>do<\/em> \u2014 the <em>actions<\/em> they perform. Example: &#8220;Systems related to access to benefits or services or assignment of penalties, such as systems that support decision-makers who adjudicate benefits &#8230;, systems which similarly assist in the adjudication of administrative or criminal penalties &#8230;&#8221;<\/p>\n\n\n\n<p><strong>Five broad rights are described. <\/strong>Copied straight from the blueprint document, with a couple of commas and boldface added:<\/p>\n\n\n\n<ol class=\"wp-block-list\"><li>&#8220;You should be protected from <strong>unsafe<\/strong> or <strong>ineffective<\/strong> systems.&#8221;<\/li><li>&#8220;You should not face <strong>discrimination<\/strong> by algorithms, and systems should be used and designed in an <strong>equitable<\/strong> way.&#8221;<\/li><li>&#8220;You should be protected from <strong>abusive data practices<\/strong> via built-in protections, and you should have <strong>agency<\/strong> over how data about you is used.&#8221;<\/li><li>&#8220;You should <strong>know<\/strong> that an automated system is being used and <strong>understand<\/strong> how and why it contributes to <strong>outcomes<\/strong> that impact you.&#8221;<\/li><li>&#8220;You should be <strong>able to opt out,<\/strong> where appropriate, and have <strong>access to a person<\/strong> who can quickly consider and remedy problems you encounter.&#8221;<\/li><\/ol>\n\n\n\n<p>I admire how plainspoken these statements are. I also feel a bit hopeless, reading them \u2014 these genies are well out of their bottles already, and I doubt any of these can ever be enforced to a meaningful degree.<\/p>\n\n\n\n<p>Just take, for example, &#8220;You should know that an automated system is being used.&#8221; Companies such as Facebook will write this into their 200,000-word terms of service, to which you must agree before signing in, and use that as evidence that &#8220;you knew.&#8221; Did you know Facebook was <a rel=\"noreferrer noopener\" href=\"https:\/\/www.theguardian.com\/technology\/2021\/feb\/04\/facebook-groups-misinformation\" target=\"_blank\">deliberately steering you and your loved ones to toxic hate groups<\/a> on the platform? No. Did you know <a rel=\"noreferrer noopener\" href=\"https:\/\/www.nytimes.com\/2021\/11\/02\/technology\/facebook-facial-recognition.html\" target=\"_blank\">your family photos were being used to train face-recognition systems<\/a>? No. Is Facebook going to give you a big, easy-to-read warning about the next invasive or exploitative technology it deploys against you? Certainly not.<\/p>\n\n\n\n<p>What about &#8220;You should be protected from abusive data practices&#8221;? For over 20 years, an algorithm ensured that Black Americans \u2014 <em>specifically<\/em> Black Americans \u2014 were <a rel=\"noreferrer noopener\" href=\"https:\/\/www.washingtonpost.com\/health\/black-kidney-patients-racial-health-disparities\/2021\/06\/04\/7752b492-c3a7-11eb-9a8d-f95d7724967c_story.html\" target=\"_blank\">recorded as having healthier kidneys than they actually had<\/a>, which meant life-saving care for many of them was delayed or withheld. (The National Kidney Foundation <a rel=\"noreferrer noopener\" href=\"https:\/\/www.kidney.org\/news\/nkf-and-asn-release-new-way-to-diagnose-kidney-diseases\" target=\"_blank\">finally addressed this in late 2021<\/a>.) Note, that isn&#8217;t even AI per se \u2014 it&#8217;s just the way authorities manipulate data for the sake of convenience, efficiency, or profit.<\/p>\n\n\n\n<p>One thing <em>missing<\/em> is the idea that you should be able to <strong>challenge<\/strong> the outcome that came from an algorithm. This might be assumed part of &#8220;understand how and why [an automated system] contributes to outcomes that impact you,&#8221; but I think it needs to be more explicit. If you are denied a bank loan, for example, you should be told <em>which variable<\/em> or variables caused the denial. Was it the house&#8217;s zip code, for example? Was it your income? What are your options to improve the outcome?<\/p>\n\n\n\n<p>You should be able to demand a test \u2014 say, running a mortgage application that is identical to yours <em>except for a few selected data points<\/em> (which might be related to, for example, your race or ethnicity). If that fictitious application is approved, it shows that your denial was unfair.<\/p>\n\n\n\n<p><strong>Enforcement<\/strong> of the five points in the blueprint is bound to be difficult, as can be seen from these few examples.<\/p>\n\n\n\n<p>.<\/p>\n\n\n\n<p><a rel=\"license\" href=\"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/\"><img decoding=\"async\" alt=\"Creative Commons License\" style=\"border-width:0\" src=\"https:\/\/i.creativecommons.org\/l\/by-nc-nd\/4.0\/88x31.png\"><\/a><br><small><span xmlns:dct=\"http:\/\/purl.org\/dc\/terms\/\" property=\"dct:title\"><strong>AI in Media and Society<\/strong><\/span> by <span xmlns:cc=\"http:\/\/creativecommons.org\/ns#\" property=\"cc:attributionName\">Mindy McAdams<\/span> is licensed under a <a rel=\"license\" href=\"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/\">Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License<\/a>.<br>Include the author&#8217;s name (Mindy McAdams) and a link to the original post in any reuse of this content.<\/small><\/p>\n\n\n\n<p>.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The White House announced a Blueprint for an AI Bill of Rights on Oct. 4. The MIT Technology Review wrote about it the same day (so did many other tech publishers). According to writer Melissa Heikkil\u00e4, &#8220;Critics say the plan lacks teeth and the US needs even tougher regulation around AI.&#8221; An associated document, titled&hellip; <a class=\"more-link\" href=\"https:\/\/www.macloo.com\/ai\/2022\/10\/07\/ai-bill-of-rights-shows-good-intentions\/\">Continue reading <span class=\"screen-reader-text\">AI Bill of Rights shows good intentions<\/span> <span class=\"meta-nav\" aria-hidden=\"true\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[217],"tags":[220,221,222,110,218],"class_list":["post-1280","post","type-post","status-publish","format-standard","hentry","category-regulation","tag-abuses","tag-accountability","tag-data","tag-government","tag-protection"],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/posts\/1280","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/comments?post=1280"}],"version-history":[{"count":10,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/posts\/1280\/revisions"}],"predecessor-version":[{"id":1292,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/posts\/1280\/revisions\/1292"}],"wp:attachment":[{"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/media?parent=1280"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/categories?post=1280"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.macloo.com\/ai\/wp-json\/wp\/v2\/tags?post=1280"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}