Actionable Auditing Revisited--Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products.

Saved in:
Bibliographic Details
Title: Actionable Auditing Revisited--Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products.
Authors: Raji, Inioluwa Deborah, Buolamwini, Joy
Source: Communications of the ACM; Jan2023, Vol. 66 Issue 1, p101-108, 8p, 1 Diagram, 2 Charts
Subject Terms: HUMAN facial recognition software, AUDITING, HUMAN skin color, GENDER, ERRORS, SOCIAL responsibility of business
Abstract: Although algorithmic auditing has emerged as a key strategy to expose systematic biases embedded in software platforms, we struggle to understand the real-world impact of these audits and continue to find it difficult to translate such independent assessments into meaningful corporate accountability. To analyze the impact of publicly naming and disclosing performance results of biased AI systems, we investigate the commercial impact of Gender Shades, the first algorithmic audit of gender- and skin-type performance disparities in commercial facial analysis models. This paper (1) outlines the audit design and structured disclosure procedure used in the Gender Shades study, (2) presents new performance metrics from targeted companies such as IBM, Microsoft, and Megvii (Face++) on the Pilot Parliaments Benchmark (PPB) as of August 2018, (3) provides performance results on PPB by non-target companies such as Amazon and Kairos, and (4) explores differences in company responses as shared through corporate communications that contextualize differences in performance on PPB. Within 7 months of the original audit, we find that all three targets released new application program interface (API) versions. All targets reduced accuracy disparities between males and females and darker- and lighter-skinned subgroups, with the most significant update occurring for the darker-skinned female subgroup that underwent a 17.7-30.4% reduction in error between audit periods. Minimizing these disparities led to a 5.72-8.3% reduction in overall error on the Pilot Parliaments Benchmark (PPB) for target corporation APIs. The overall performance of non-targets Amazon and Kairos lags significantly behind that of the targets, with error rates of 8.66% and 6.60% overall, and error rates of 31.37% and 22.50% for the darker female subgroup, respectively. This is an expanded version of an earlier publication of these results, revised for a more general audience, and updated to include commentary on further developments. [ABSTRACT FROM AUTHOR]
Copyright of Communications of the ACM is the property of Association for Computing Machinery and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
Database: Complementary Index
FullText Text:
  Availability: 0
CustomLinks:
  – Url: https://resolver.ebscohost.com/openurl?sid=EBSCO:edb&genre=article&issn=00010782&ISBN=&volume=66&issue=1&date=20230101&spage=101&pages=101-108&title=Communications of the ACM&atitle=Actionable%20Auditing%20Revisited--Investigating%20the%20Impact%20of%20Publicly%20Naming%20Biased%20Performance%20Results%20of%20Commercial%20AI%20Products.&aulast=Raji%2C%20Inioluwa%20Deborah&id=DOI:10.1145/3571151
    Name: Full Text Finder
    Category: fullText
    Text: Full Text Finder
    Icon: https://imageserver.ebscohost.com/branding/images/FTF.gif
    MouseOverText: Full Text Finder
  – Url: https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=EBSCO&SrcAuth=EBSCO&DestApp=WOS&ServiceName=TransferToWoS&DestLinkType=GeneralSearchSummary&Func=Links&author=Raji%20ID
    Name: ISI
    Category: fullText
    Text: Nájsť tento článok vo Web of Science
    Icon: https://imagesrvr.epnet.com/ls/20docs.gif
    MouseOverText: Nájsť tento článok vo Web of Science
Header DbId: edb
DbLabel: Complementary Index
An: 160917440
RelevancyScore: 1003
AccessLevel: 6
PubType: Periodical
PubTypeId: serialPeriodical
PreciseRelevancyScore: 1002.84948730469
IllustrationInfo
Items – Name: Title
  Label: Title
  Group: Ti
  Data: Actionable Auditing Revisited--Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products.
– Name: Author
  Label: Authors
  Group: Au
  Data: <searchLink fieldCode="AR" term="%22Raji%2C+Inioluwa+Deborah%22">Raji, Inioluwa Deborah</searchLink><br /><searchLink fieldCode="AR" term="%22Buolamwini%2C+Joy%22">Buolamwini, Joy</searchLink>
– Name: TitleSource
  Label: Source
  Group: Src
  Data: Communications of the ACM; Jan2023, Vol. 66 Issue 1, p101-108, 8p, 1 Diagram, 2 Charts
– Name: Subject
  Label: Subject Terms
  Group: Su
  Data: <searchLink fieldCode="DE" term="%22HUMAN+facial+recognition+software%22">HUMAN facial recognition software</searchLink><br /><searchLink fieldCode="DE" term="%22AUDITING%22">AUDITING</searchLink><br /><searchLink fieldCode="DE" term="%22HUMAN+skin+color%22">HUMAN skin color</searchLink><br /><searchLink fieldCode="DE" term="%22GENDER%22">GENDER</searchLink><br /><searchLink fieldCode="DE" term="%22ERRORS%22">ERRORS</searchLink><br /><searchLink fieldCode="DE" term="%22SOCIAL+responsibility+of+business%22">SOCIAL responsibility of business</searchLink>
– Name: Abstract
  Label: Abstract
  Group: Ab
  Data: Although algorithmic auditing has emerged as a key strategy to expose systematic biases embedded in software platforms, we struggle to understand the real-world impact of these audits and continue to find it difficult to translate such independent assessments into meaningful corporate accountability. To analyze the impact of publicly naming and disclosing performance results of biased AI systems, we investigate the commercial impact of Gender Shades, the first algorithmic audit of gender- and skin-type performance disparities in commercial facial analysis models. This paper (1) outlines the audit design and structured disclosure procedure used in the Gender Shades study, (2) presents new performance metrics from targeted companies such as IBM, Microsoft, and Megvii (Face++) on the Pilot Parliaments Benchmark (PPB) as of August 2018, (3) provides performance results on PPB by non-target companies such as Amazon and Kairos, and (4) explores differences in company responses as shared through corporate communications that contextualize differences in performance on PPB. Within 7 months of the original audit, we find that all three targets released new application program interface (API) versions. All targets reduced accuracy disparities between males and females and darker- and lighter-skinned subgroups, with the most significant update occurring for the darker-skinned female subgroup that underwent a 17.7-30.4% reduction in error between audit periods. Minimizing these disparities led to a 5.72-8.3% reduction in overall error on the Pilot Parliaments Benchmark (PPB) for target corporation APIs. The overall performance of non-targets Amazon and Kairos lags significantly behind that of the targets, with error rates of 8.66% and 6.60% overall, and error rates of 31.37% and 22.50% for the darker female subgroup, respectively. This is an expanded version of an earlier publication of these results, revised for a more general audience, and updated to include commentary on further developments. [ABSTRACT FROM AUTHOR]
– Name: Abstract
  Label:
  Group: Ab
  Data: <i>Copyright of Communications of the ACM is the property of Association for Computing Machinery and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract.</i> (Copyright applies to all Abstracts.)
PLink https://erproxy.cvtisr.sk/sfx/access?url=https://search.ebscohost.com/login.aspx?direct=true&site=eds-live&db=edb&AN=160917440
RecordInfo BibRecord:
  BibEntity:
    Identifiers:
      – Type: doi
        Value: 10.1145/3571151
    Languages:
      – Code: eng
        Text: English
    PhysicalDescription:
      Pagination:
        PageCount: 8
        StartPage: 101
    Subjects:
      – SubjectFull: HUMAN facial recognition software
        Type: general
      – SubjectFull: AUDITING
        Type: general
      – SubjectFull: HUMAN skin color
        Type: general
      – SubjectFull: GENDER
        Type: general
      – SubjectFull: ERRORS
        Type: general
      – SubjectFull: SOCIAL responsibility of business
        Type: general
    Titles:
      – TitleFull: Actionable Auditing Revisited--Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products.
        Type: main
  BibRelationships:
    HasContributorRelationships:
      – PersonEntity:
          Name:
            NameFull: Raji, Inioluwa Deborah
      – PersonEntity:
          Name:
            NameFull: Buolamwini, Joy
    IsPartOfRelationships:
      – BibEntity:
          Dates:
            – D: 01
              M: 01
              Text: Jan2023
              Type: published
              Y: 2023
          Identifiers:
            – Type: issn-print
              Value: 00010782
          Numbering:
            – Type: volume
              Value: 66
            – Type: issue
              Value: 1
          Titles:
            – TitleFull: Communications of the ACM
              Type: main
ResultId 1