Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings

Abstract This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing...

Full description

Bibliographic Details
Main Authors: Georgios Ioannides, Aishwarya Jadhav, Aditi Sharma, Samarth Navali, Alan W. Black
Format: Article
Language:English
Published: Nature Portfolio 2023-10-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-023-45677-0
_version_ 1797647343498035200
author Georgios Ioannides
Aishwarya Jadhav
Aditi Sharma
Samarth Navali
Alan W. Black
author_facet Georgios Ioannides
Aishwarya Jadhav
Aditi Sharma
Samarth Navali
Alan W. Black
author_sort Georgios Ioannides
collection DOAJ
description Abstract This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing or reducing the gender component from the embeddings of target words, while preserving useful semantic information. Their gender bias is assessed through the Word Embedding Association Test. The performance of co-reference resolution and text classification models trained on both original and debiased embeddings is evaluated in terms of accuracy. A compressed co-reference resolution model is examined to gauge the effectiveness of debiasing techniques on resource-efficient models. To the best of the authors’ knowledge, this is the first attempt to apply compression techniques to debiased models. By analyzing the context preservation of debiased embeddings using a Twitter misinformation dataset, this study contributes valuable insights into the practical implications of debiasing methods for real-world applications such as person profiling.
first_indexed 2024-03-11T15:14:54Z
format Article
id doaj.art-87033d2a03fd4e73b5043d2b0f03e7db
institution Directory Open Access Journal
issn 2045-2322
language English
last_indexed 2024-03-11T15:14:54Z
publishDate 2023-10-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj.art-87033d2a03fd4e73b5043d2b0f03e7db2023-10-29T12:22:58ZengNature PortfolioScientific Reports2045-23222023-10-011311910.1038/s41598-023-45677-0Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddingsGeorgios Ioannides0Aishwarya Jadhav1Aditi Sharma2Samarth Navali3Alan W. Black4Language Technologies Institute, Carnegie Mellon UniversityLanguage Technologies Institute, Carnegie Mellon UniversityLanguage Technologies Institute, Carnegie Mellon UniversityLanguage Technologies Institute, Carnegie Mellon UniversityLanguage Technologies Institute, Carnegie Mellon UniversityAbstract This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing or reducing the gender component from the embeddings of target words, while preserving useful semantic information. Their gender bias is assessed through the Word Embedding Association Test. The performance of co-reference resolution and text classification models trained on both original and debiased embeddings is evaluated in terms of accuracy. A compressed co-reference resolution model is examined to gauge the effectiveness of debiasing techniques on resource-efficient models. To the best of the authors’ knowledge, this is the first attempt to apply compression techniques to debiased models. By analyzing the context preservation of debiased embeddings using a Twitter misinformation dataset, this study contributes valuable insights into the practical implications of debiasing methods for real-world applications such as person profiling.https://doi.org/10.1038/s41598-023-45677-0
spellingShingle Georgios Ioannides
Aishwarya Jadhav
Aditi Sharma
Samarth Navali
Alan W. Black
Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
Scientific Reports
title Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_full Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_fullStr Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_full_unstemmed Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_short Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_sort compressed models for co reference resolution enhancing efficiency with debiased word embeddings
url https://doi.org/10.1038/s41598-023-45677-0
work_keys_str_mv AT georgiosioannides compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT aishwaryajadhav compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT aditisharma compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT samarthnavali compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT alanwblack compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings