Mixup Based Cross-Consistency Training for Named Entity Recognition

Named Entity Recognition (NER) is at the core of natural language understanding. The quality and amount of datasets determine the performance of deep-learning-based NER models. As datasets for NER require token-level or word-level labels to be assigned, annotating the datasets is expensive and time...

Full description

Bibliographic Details
Main Authors: Geonsik Youn, Bohan Yoon, Seungbin Ji, Dahee Ko, Jongtae Rhee
Format: Article
Language:English
Published: MDPI AG 2022-11-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/12/21/11084
_version_ 1797469230576173056
author Geonsik Youn
Bohan Yoon
Seungbin Ji
Dahee Ko
Jongtae Rhee
author_facet Geonsik Youn
Bohan Yoon
Seungbin Ji
Dahee Ko
Jongtae Rhee
author_sort Geonsik Youn
collection DOAJ
description Named Entity Recognition (NER) is at the core of natural language understanding. The quality and amount of datasets determine the performance of deep-learning-based NER models. As datasets for NER require token-level or word-level labels to be assigned, annotating the datasets is expensive and time consuming. To alleviate efforts of manual anotation, many prior studies utilized weak supervision for NER tasks. However, using weak supervision directly would be an obstacle for training deep networks because the labels automatically annotated contain a a lot of noise. In this study, we propose a framework to better train the deep model for NER tasks using weakly labeled data. The proposed framework stems from the idea that mixup, which was recently considered as a data augmentation strategy, would be an obstacle to deep model training for NER tasks. Inspired by this idea, we used mixup as a perturbation function for consistency regularization, one of the semi-supervised learning strategies. To support our idea, we conducted several experiments for NER benchmarks. Experimental results proved that directly using mixup on NER tasks hinders deep model training while demonstrating that the proposed framework achieves improved performances compared to employing only a few human-annotated data.
first_indexed 2024-03-09T19:18:26Z
format Article
id doaj.art-b9f8d9d68f064c22a7899fb170ee5939
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-09T19:18:26Z
publishDate 2022-11-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-b9f8d9d68f064c22a7899fb170ee59392023-11-24T03:37:47ZengMDPI AGApplied Sciences2076-34172022-11-0112211108410.3390/app122111084Mixup Based Cross-Consistency Training for Named Entity RecognitionGeonsik Youn0Bohan Yoon1Seungbin Ji2Dahee Ko3Jongtae Rhee4Department of Industrial and Systems Engineering, Dongguk University, Seoul 04620, KoreaDepartment of Industrial and Systems Engineering, Dongguk University, Seoul 04620, KoreaDepartment of Industrial and Systems Engineering, Dongguk University, Seoul 04620, KoreaDepartment of Industrial and Systems Engineering, Dongguk University, Seoul 04620, KoreaDepartment of Industrial and Systems Engineering, Dongguk University, Seoul 04620, KoreaNamed Entity Recognition (NER) is at the core of natural language understanding. The quality and amount of datasets determine the performance of deep-learning-based NER models. As datasets for NER require token-level or word-level labels to be assigned, annotating the datasets is expensive and time consuming. To alleviate efforts of manual anotation, many prior studies utilized weak supervision for NER tasks. However, using weak supervision directly would be an obstacle for training deep networks because the labels automatically annotated contain a a lot of noise. In this study, we propose a framework to better train the deep model for NER tasks using weakly labeled data. The proposed framework stems from the idea that mixup, which was recently considered as a data augmentation strategy, would be an obstacle to deep model training for NER tasks. Inspired by this idea, we used mixup as a perturbation function for consistency regularization, one of the semi-supervised learning strategies. To support our idea, we conducted several experiments for NER benchmarks. Experimental results proved that directly using mixup on NER tasks hinders deep model training while demonstrating that the proposed framework achieves improved performances compared to employing only a few human-annotated data.https://www.mdpi.com/2076-3417/12/21/11084deep learningnamed entity recognitionconsistency regularizationsemi-supervised learningmixup
spellingShingle Geonsik Youn
Bohan Yoon
Seungbin Ji
Dahee Ko
Jongtae Rhee
Mixup Based Cross-Consistency Training for Named Entity Recognition
Applied Sciences
deep learning
named entity recognition
consistency regularization
semi-supervised learning
mixup
title Mixup Based Cross-Consistency Training for Named Entity Recognition
title_full Mixup Based Cross-Consistency Training for Named Entity Recognition
title_fullStr Mixup Based Cross-Consistency Training for Named Entity Recognition
title_full_unstemmed Mixup Based Cross-Consistency Training for Named Entity Recognition
title_short Mixup Based Cross-Consistency Training for Named Entity Recognition
title_sort mixup based cross consistency training for named entity recognition
topic deep learning
named entity recognition
consistency regularization
semi-supervised learning
mixup
url https://www.mdpi.com/2076-3417/12/21/11084
work_keys_str_mv AT geonsikyoun mixupbasedcrossconsistencytrainingfornamedentityrecognition
AT bohanyoon mixupbasedcrossconsistencytrainingfornamedentityrecognition
AT seungbinji mixupbasedcrossconsistencytrainingfornamedentityrecognition
AT daheeko mixupbasedcrossconsistencytrainingfornamedentityrecognition
AT jongtaerhee mixupbasedcrossconsistencytrainingfornamedentityrecognition