Reliance on metrics is a fundamental challenge for AI
Summary: Through a series of case studies, we review how the unthinking pursuit of metric optimization can lead to real-world harms, including recommendation systems promoting radicalization, well-loved teachers fired by an algorithm, and essay grading software that rewards sophisticated garbage. Th...
| Published in: | Patterns |
|---|---|
| Main Authors: | , |
| Format: | Article |
| Language: | English |
| Published: |
Elsevier
2022-05-01
|
| Subjects: | |
| Online Access: | http://www.sciencedirect.com/science/article/pii/S2666389922000563 |
| _version_ | 1852656272393895936 |
|---|---|
| author | Rachel L. Thomas David Uminsky |
| author_facet | Rachel L. Thomas David Uminsky |
| author_sort | Rachel L. Thomas |
| collection | DOAJ |
| container_title | Patterns |
| description | Summary: Through a series of case studies, we review how the unthinking pursuit of metric optimization can lead to real-world harms, including recommendation systems promoting radicalization, well-loved teachers fired by an algorithm, and essay grading software that rewards sophisticated garbage. The metrics used are often proxies for underlying, unmeasurable quantities (e.g., “watch time” of a video as a proxy for “user satisfaction”). We propose an evidence-based framework to mitigate such harms by (1) using a slate of metrics to get a fuller and more nuanced picture; (2) conducting external algorithmic audits; (3) combining metrics with qualitative accounts; and (4) involving a range of stakeholders, including those who will be most impacted. The bigger picture: The success of current artificial intelligence (AI) approaches such as deep learning centers on their unreasonable effectiveness at metric optimization, yet overemphasizing metrics leads to a variety of real-world harms, including manipulation, gaming, and a myopic focus on short-term qualities and inadequate proxies. This principle is classically captured in Goodhart’s law: when a measure becomes the target, it ceases to be an effective measure. Current AI approaches have we aponized Goodhart’s law by centering on optimizing a particular measure as a target. This poses a grand contradiction within AI design and ethics: optimizing metrics results in far from optimal outcomes. It is crucial to understand this dynamic in order to mitigate the risks and harms we are facing as a result of misuse of AI. |
| format | Article |
| id | doaj-art-e0e6f9f27136434fbd0d7dc2898f4e3b |
| institution | Directory of Open Access Journals |
| issn | 2666-3899 |
| language | English |
| publishDate | 2022-05-01 |
| publisher | Elsevier |
| record_format | Article |
| spelling | doaj-art-e0e6f9f27136434fbd0d7dc2898f4e3b2025-08-19T21:38:46ZengElsevierPatterns2666-38992022-05-013510047610.1016/j.patter.2022.100476Reliance on metrics is a fundamental challenge for AIRachel L. Thomas0David Uminsky1Queensland University of Technology, Brisbane, QLD, AustraliaUniversity of Chicago, Chicago, IL, USA; Corresponding authorSummary: Through a series of case studies, we review how the unthinking pursuit of metric optimization can lead to real-world harms, including recommendation systems promoting radicalization, well-loved teachers fired by an algorithm, and essay grading software that rewards sophisticated garbage. The metrics used are often proxies for underlying, unmeasurable quantities (e.g., “watch time” of a video as a proxy for “user satisfaction”). We propose an evidence-based framework to mitigate such harms by (1) using a slate of metrics to get a fuller and more nuanced picture; (2) conducting external algorithmic audits; (3) combining metrics with qualitative accounts; and (4) involving a range of stakeholders, including those who will be most impacted. The bigger picture: The success of current artificial intelligence (AI) approaches such as deep learning centers on their unreasonable effectiveness at metric optimization, yet overemphasizing metrics leads to a variety of real-world harms, including manipulation, gaming, and a myopic focus on short-term qualities and inadequate proxies. This principle is classically captured in Goodhart’s law: when a measure becomes the target, it ceases to be an effective measure. Current AI approaches have we aponized Goodhart’s law by centering on optimizing a particular measure as a target. This poses a grand contradiction within AI design and ethics: optimizing metrics results in far from optimal outcomes. It is crucial to understand this dynamic in order to mitigate the risks and harms we are facing as a result of misuse of AI.http://www.sciencedirect.com/science/article/pii/S2666389922000563DSML 1: Concept: Basic principles of a new data science output observed and reported |
| spellingShingle | Rachel L. Thomas David Uminsky Reliance on metrics is a fundamental challenge for AI DSML 1: Concept: Basic principles of a new data science output observed and reported |
| title | Reliance on metrics is a fundamental challenge for AI |
| title_full | Reliance on metrics is a fundamental challenge for AI |
| title_fullStr | Reliance on metrics is a fundamental challenge for AI |
| title_full_unstemmed | Reliance on metrics is a fundamental challenge for AI |
| title_short | Reliance on metrics is a fundamental challenge for AI |
| title_sort | reliance on metrics is a fundamental challenge for ai |
| topic | DSML 1: Concept: Basic principles of a new data science output observed and reported |
| url | http://www.sciencedirect.com/science/article/pii/S2666389922000563 |
| work_keys_str_mv | AT rachellthomas relianceonmetricsisafundamentalchallengeforai AT daviduminsky relianceonmetricsisafundamentalchallengeforai |
