Empowering Community-Driven Determination of Values for Language Models
Emerging technologies like Artificial Intelligence and Large Language Models are often developed in Western contexts and carry implicit values, from developer choices or underlying training data, which are not adequately representative of the diverse contexts in which they are deployed. The resultan...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2024
|
Online Access: | https://hdl.handle.net/1721.1/157021 |
_version_ | 1824458428740796416 |
---|---|
author | Raman, Deepika |
author2 | Hadfield-Menell, Dylan |
author_facet | Hadfield-Menell, Dylan Raman, Deepika |
author_sort | Raman, Deepika |
collection | MIT |
description | Emerging technologies like Artificial Intelligence and Large Language Models are often developed in Western contexts and carry implicit values, from developer choices or underlying training data, which are not adequately representative of the diverse contexts in which they are deployed. The resultant misalignment from the lack of engagement with non-Eurocentric value paradigms results in inadequate, and potentially harmful outcomes that impact these unconsidered communities. To codify fundamentally subjective human values therefore necessitates the elicitation of these nuances through the inclusion and involvement of these very communities.
This thesis argues that participants’ lack of familiarity with new technologies like Artificial Intelligence impacts their engagement and contribution to participatory processes of AI development. This thesis also helps demonstrate how grounded theory approaches can be leveraged to contextualize awareness-building efforts that can potentially empower community participation by addressing such familiarity gaps.
This two-fold objective of (i)eliciting community-relevant attributes for language model alignment (ii)through the necessary familiarization of the technology in question is demonstrated through the means of sample case studies. A grounded participatory process CALMA (Community-aligned Axes for Language Model Alignment) is designed and evaluated through these cases to illustrate this contextualized alignment exercise. Learnings from this comparative case study are then extended to explore avenues for communities and institutions to adopt similar techniques that center the voices of the final users. |
first_indexed | 2025-02-19T04:25:44Z |
format | Thesis |
id | mit-1721.1/157021 |
institution | Massachusetts Institute of Technology |
last_indexed | 2025-02-19T04:25:44Z |
publishDate | 2024 |
publisher | Massachusetts Institute of Technology |
record_format | dspace |
spelling | mit-1721.1/1570212024-09-25T03:02:50Z Empowering Community-Driven Determination of Values for Language Models Raman, Deepika Hadfield-Menell, Dylan Massachusetts Institute of Technology. Institute for Data, Systems, and Society Technology and Policy Program Emerging technologies like Artificial Intelligence and Large Language Models are often developed in Western contexts and carry implicit values, from developer choices or underlying training data, which are not adequately representative of the diverse contexts in which they are deployed. The resultant misalignment from the lack of engagement with non-Eurocentric value paradigms results in inadequate, and potentially harmful outcomes that impact these unconsidered communities. To codify fundamentally subjective human values therefore necessitates the elicitation of these nuances through the inclusion and involvement of these very communities. This thesis argues that participants’ lack of familiarity with new technologies like Artificial Intelligence impacts their engagement and contribution to participatory processes of AI development. This thesis also helps demonstrate how grounded theory approaches can be leveraged to contextualize awareness-building efforts that can potentially empower community participation by addressing such familiarity gaps. This two-fold objective of (i)eliciting community-relevant attributes for language model alignment (ii)through the necessary familiarization of the technology in question is demonstrated through the means of sample case studies. A grounded participatory process CALMA (Community-aligned Axes for Language Model Alignment) is designed and evaluated through these cases to illustrate this contextualized alignment exercise. Learnings from this comparative case study are then extended to explore avenues for communities and institutions to adopt similar techniques that center the voices of the final users. S.M. 2024-09-24T18:27:13Z 2024-09-24T18:27:13Z 2024-05 2024-07-25T14:17:40.697Z Thesis https://hdl.handle.net/1721.1/157021 0000-0001-6726-5482 Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) Copyright retained by author(s) https://creativecommons.org/licenses/by-nc-nd/4.0/ application/pdf Massachusetts Institute of Technology |
spellingShingle | Raman, Deepika Empowering Community-Driven Determination of Values for Language Models |
title | Empowering Community-Driven Determination of Values for Language Models |
title_full | Empowering Community-Driven Determination of Values for Language Models |
title_fullStr | Empowering Community-Driven Determination of Values for Language Models |
title_full_unstemmed | Empowering Community-Driven Determination of Values for Language Models |
title_short | Empowering Community-Driven Determination of Values for Language Models |
title_sort | empowering community driven determination of values for language models |
url | https://hdl.handle.net/1721.1/157021 |
work_keys_str_mv | AT ramandeepika empoweringcommunitydrivendeterminationofvaluesforlanguagemodels |