Context and Participation in Machine Learning

ML systems are shaped by human choices and norms, from problem conceptualization to deployment. They are then used in complex socio-technical contexts, where they interact with and affect diverse populations. However, development decisions are often made in isolation, without deeply taking into ac...

Full description

Bibliographic Details
Main Author: Suresh, Harini
Other Authors: Guttag, John V.
Format: Thesis
Published: Massachusetts Institute of Technology 2023
Online Access:https://hdl.handle.net/1721.1/150314
_version_ 1826199898531299328
author Suresh, Harini
author2 Guttag, John V.
author_facet Guttag, John V.
Suresh, Harini
author_sort Suresh, Harini
collection MIT
description ML systems are shaped by human choices and norms, from problem conceptualization to deployment. They are then used in complex socio-technical contexts, where they interact with and affect diverse populations. However, development decisions are often made in isolation, without deeply taking into account the deployment context in which the system will be used. And they are typically hidden to users in that context, who have few avenues to understand if or how they should use the system. As a result, there are numerous examples of ML systems that in practice are harmful, poorly understood, or misused. We propose an alternate approach to the development and deployment of ML systems that is focused on incorporating the participation of the people who use and are affected by the system. We first develop two frameworks that lend clarity to the human choices that shape ML systems and the broad populations that these systems affect. These inform a prospective question: how can we shape new systems from the start to reflect context-specific needs and benefit justice and equity? We address this question through an in-depth case study of co-designing ML tools to support activists who monitor gender-related violence. Drawing from intersectional feminist theory and participatory design, we develop methods for data collection, annotation, modeling, and evaluation that prioritize sustainable partnerships and challenge power inequalities. Then, we consider an alternative paradigm where we do not have full control over the development lifecycle, e.g., where a model has already been built and made available. In these cases, we show how deployment tools can give downstream stakeholders the information and agency to understand and hold ML systems accountable. We describe the design of two novel deployment tools that provide intuitive, useful, and context-relevant insight into model strengths and limitations. The first uses example-based visualizations and an interactive input editor to help users assess the reliability of individual model predictions. The second, Kaleidoscope, enables context-specific evaluation, allowing downstream users to translate their implicit knowledge of "good model behavior'' for their context into explicitly-defined, semantically-meaningful tests. This dissertation demonstrates several ways that context-specific considerations and meaningful participation can shape the development and use of ML systems. We hope that this is a step towards the broader goal of building ML-based systems that are grounded in societal context, are shaped by diverse viewpoints, and contribute to justice and equity.
first_indexed 2024-09-23T11:27:52Z
format Thesis
id mit-1721.1/150314
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T11:27:52Z
publishDate 2023
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1503142023-04-01T03:39:06Z Context and Participation in Machine Learning Suresh, Harini Guttag, John V. Satyanarayan, Arvind Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science ML systems are shaped by human choices and norms, from problem conceptualization to deployment. They are then used in complex socio-technical contexts, where they interact with and affect diverse populations. However, development decisions are often made in isolation, without deeply taking into account the deployment context in which the system will be used. And they are typically hidden to users in that context, who have few avenues to understand if or how they should use the system. As a result, there are numerous examples of ML systems that in practice are harmful, poorly understood, or misused. We propose an alternate approach to the development and deployment of ML systems that is focused on incorporating the participation of the people who use and are affected by the system. We first develop two frameworks that lend clarity to the human choices that shape ML systems and the broad populations that these systems affect. These inform a prospective question: how can we shape new systems from the start to reflect context-specific needs and benefit justice and equity? We address this question through an in-depth case study of co-designing ML tools to support activists who monitor gender-related violence. Drawing from intersectional feminist theory and participatory design, we develop methods for data collection, annotation, modeling, and evaluation that prioritize sustainable partnerships and challenge power inequalities. Then, we consider an alternative paradigm where we do not have full control over the development lifecycle, e.g., where a model has already been built and made available. In these cases, we show how deployment tools can give downstream stakeholders the information and agency to understand and hold ML systems accountable. We describe the design of two novel deployment tools that provide intuitive, useful, and context-relevant insight into model strengths and limitations. The first uses example-based visualizations and an interactive input editor to help users assess the reliability of individual model predictions. The second, Kaleidoscope, enables context-specific evaluation, allowing downstream users to translate their implicit knowledge of "good model behavior'' for their context into explicitly-defined, semantically-meaningful tests. This dissertation demonstrates several ways that context-specific considerations and meaningful participation can shape the development and use of ML systems. We hope that this is a step towards the broader goal of building ML-based systems that are grounded in societal context, are shaped by diverse viewpoints, and contribute to justice and equity. Ph.D. 2023-03-31T14:47:06Z 2023-03-31T14:47:06Z 2023-02 2023-02-28T14:39:15.867Z Thesis https://hdl.handle.net/1721.1/150314 0000-0002-9769-4947 In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Suresh, Harini
Context and Participation in Machine Learning
title Context and Participation in Machine Learning
title_full Context and Participation in Machine Learning
title_fullStr Context and Participation in Machine Learning
title_full_unstemmed Context and Participation in Machine Learning
title_short Context and Participation in Machine Learning
title_sort context and participation in machine learning
url https://hdl.handle.net/1721.1/150314
work_keys_str_mv AT sureshharini contextandparticipationinmachinelearning