VPN: Variation on Prompt Tuning for Named-Entity Recognition

Recently, prompt-based methods have achieved a promising performance in many natural language processing benchmarks. Despite success in sentence-level classification tasks, prompt-based methods work poorly in token-level tasks, such as named entity recognition (NER), due to the sophisticated design...

Full description

Bibliographic Details
Main Authors: Niu Hu, Xuan Zhou, Bing Xu, Hanqing Liu, Xiangjin Xie, Hai-Tao Zheng
Format: Article
Language:English
Published: MDPI AG 2023-07-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/14/8359
_version_ 1797590349865025536
author Niu Hu
Xuan Zhou
Bing Xu
Hanqing Liu
Xiangjin Xie
Hai-Tao Zheng
author_facet Niu Hu
Xuan Zhou
Bing Xu
Hanqing Liu
Xiangjin Xie
Hai-Tao Zheng
author_sort Niu Hu
collection DOAJ
description Recently, prompt-based methods have achieved a promising performance in many natural language processing benchmarks. Despite success in sentence-level classification tasks, prompt-based methods work poorly in token-level tasks, such as named entity recognition (NER), due to the sophisticated design of entity-related templates. Note that the nature of prompt tuning makes full use of the parameters of the mask language model (MLM) head, while previous methods solely utilized the last hidden layer of language models (LMs) and the power of the MLM head is overlooked. In this work, we discovered the characteristics of semantic feature changes in samples after being processed using MLMs. Based on this characteristic, we designed a prompt-tuning variant for NER tasks. We let the pre-trained model predict the label words derived from the training dataset at each position and fed the generated logits (non-normalized probability) to the CRF layer. We evaluated our method on three popular datasets, and the experiments showed that our proposed method outperforms the state-of-the-art model in all three Chinese datasets.
first_indexed 2024-03-11T01:19:13Z
format Article
id doaj.art-0e8431517258472c8df832bc73f1a8d7
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T01:19:13Z
publishDate 2023-07-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-0e8431517258472c8df832bc73f1a8d72023-11-18T18:11:58ZengMDPI AGApplied Sciences2076-34172023-07-011314835910.3390/app13148359VPN: Variation on Prompt Tuning for Named-Entity RecognitionNiu Hu0Xuan Zhou1Bing Xu2Hanqing Liu3Xiangjin Xie4Hai-Tao Zheng5Shenzhen International Graduate School, Tsinghua University, Shenzhen 518055, ChinaPAII Inc., Palo Alto, CA 94306, USAPing An Technology (Shenzhen) Co., Ltd., Shenzhen 518063, ChinaShenzhen International Graduate School, Tsinghua University, Shenzhen 518055, ChinaShenzhen International Graduate School, Tsinghua University, Shenzhen 518055, ChinaShenzhen International Graduate School, Tsinghua University, Shenzhen 518055, ChinaRecently, prompt-based methods have achieved a promising performance in many natural language processing benchmarks. Despite success in sentence-level classification tasks, prompt-based methods work poorly in token-level tasks, such as named entity recognition (NER), due to the sophisticated design of entity-related templates. Note that the nature of prompt tuning makes full use of the parameters of the mask language model (MLM) head, while previous methods solely utilized the last hidden layer of language models (LMs) and the power of the MLM head is overlooked. In this work, we discovered the characteristics of semantic feature changes in samples after being processed using MLMs. Based on this characteristic, we designed a prompt-tuning variant for NER tasks. We let the pre-trained model predict the label words derived from the training dataset at each position and fed the generated logits (non-normalized probability) to the CRF layer. We evaluated our method on three popular datasets, and the experiments showed that our proposed method outperforms the state-of-the-art model in all three Chinese datasets.https://www.mdpi.com/2076-3417/13/14/8359prompt tuningMLM headNER
spellingShingle Niu Hu
Xuan Zhou
Bing Xu
Hanqing Liu
Xiangjin Xie
Hai-Tao Zheng
VPN: Variation on Prompt Tuning for Named-Entity Recognition
Applied Sciences
prompt tuning
MLM head
NER
title VPN: Variation on Prompt Tuning for Named-Entity Recognition
title_full VPN: Variation on Prompt Tuning for Named-Entity Recognition
title_fullStr VPN: Variation on Prompt Tuning for Named-Entity Recognition
title_full_unstemmed VPN: Variation on Prompt Tuning for Named-Entity Recognition
title_short VPN: Variation on Prompt Tuning for Named-Entity Recognition
title_sort vpn variation on prompt tuning for named entity recognition
topic prompt tuning
MLM head
NER
url https://www.mdpi.com/2076-3417/13/14/8359
work_keys_str_mv AT niuhu vpnvariationonprompttuningfornamedentityrecognition
AT xuanzhou vpnvariationonprompttuningfornamedentityrecognition
AT bingxu vpnvariationonprompttuningfornamedentityrecognition
AT hanqingliu vpnvariationonprompttuningfornamedentityrecognition
AT xiangjinxie vpnvariationonprompttuningfornamedentityrecognition
AT haitaozheng vpnvariationonprompttuningfornamedentityrecognition