Leveraging linguistic knowledge to enhance low-resource NLP applications

Natural Language Processing (NLP) empowers computers to process and analyze vast amounts of text data. The introduction of pre-trained language models (PLMs) has significantly advanced NLP by incorporating deep learning algorithms, thereby enhancing the handling of natural language understanding (NL...

Full description

Bibliographic Details
Main Author: Zhu, Zixiao
Other Authors: Mao Kezhi
Format: Thesis-Doctor of Philosophy
Language:English
Published: Nanyang Technological University 2025
Subjects:
Online Access:https://hdl.handle.net/10356/182513
Description
Summary:Natural Language Processing (NLP) empowers computers to process and analyze vast amounts of text data. The introduction of pre-trained language models (PLMs) has significantly advanced NLP by incorporating deep learning algorithms, thereby enhancing the handling of natural language understanding (NLU) tasks. However, due to their universal design, PLMs might not perform optimally in specialized tasks if essential features are not included during initial training. As a result, several training paradigms have been developed to enhance the downstream performance of PLMs, as outlined below. Promptless fine-tuning, a common training method, adapts PLMs to specific tasks by modifying model parameters using task-specific training data. This approach has proven effective across different low-resource deep-learning models. Nevertheless, fine-tuning may face challenges such as overfitting or lack of robustness under conditions of training data scarcity. To mitigate this, the prompt-based learning paradigm is introduced, utilizing natural language prompts to enhance model understanding. Within this paradigm, fixed-prompt LM tuning wraps input sentences into a template, allowing the PLM to engage in tasks like masked language prediction or inference understanding. These techniques, which incorporate task descriptions, have shown substantial efficacy in few-shot learning. However, the need for highly comprehensive and generalizable templates presents challenges in design, and the issues of promptless fine-tuning persist, albeit mitigated, in low-data settings. With the rise of large language models (LLMs), traditional fine-tuning has become prohibitive for many users due to computational demands. As a result, tuning-free prompting has emerged as a novel approach for those with limited computational resources, leveraging the inherent language understanding capabilities of LLMs. This paradigm, including in-context learning (ICL), relies on few-shot instruction and demonstration to prompt LLM responses. The effectiveness of ICL heavily depends on how sample-label pairs are organized within demonstrations. The strategic selection and ranking of these pairs to maximize understanding with minimal data is a critical area of ongoing research in NLP applications. External knowledge is proven to be beneficial for deep learning algorithms to reduce the reliance on training data and provide additional useful information. How to effectively incorporate useful external knowledge to excite LM's capabilities in low-resource NLP applications remains an open research question. This thesis investigates how incorporating existing knowledge into training paradigms can enhance NLP applications under low-resource conditions such as training data scarcity and limited computational resources. By leveraging external knowledge as prior knowledge, we aim to achieve improved text representations, more nuanced task descriptions, and richer label information instructions, thereby reducing the model's dependency on training data and enhancing its understanding capabilities. We demonstrate the advantages of integrating additional knowledge into deep learning systems and offer frameworks to apply this knowledge across different training paradigms, thereby improving performance on various NLP tasks, particularly under low-resource conditions. Specifically, 1. In the promptless fine-tuning paradigm, we first focus on fine-tuning word embeddings for task-related words, thereby enriching the conceptual knowledge available to compositional neural networks during feature learning in emotion recognition. This approach effectively enhances emotional keyword attention. We then extend this method by incorporating domain-specific lexical knowledge to improve the pre-trained word representations within a learning network, enriching the context-based word embeddings with discriminative features, providing more semantic insights, and bolstering performance across various classification tasks. 2. In the fixed-prompt LM tuning paradigm, we introduce a novel task description that incorporates dictionary knowledge to offer extensive semantic insights into labels. Building on this strategy, we devise an approach to augment few-shot classification performance within an entailment-based framework, significantly enhancing the efficiency of using limited training data and even facilitating zero-shot learning. 3. In the tuning-free prompting paradigm, we demonstrate how to incorporate label-related words into demonstrations based on LLM feedback, creating effective sample-and-label-level demonstrations. Additionally, we propose an innovative method that uses multiple-label words in demonstrations instead of traditional class names, offering more detailed and varied label instructions for LM understanding, thereby improving in-context learning (ICL) classification capabilities.