What does Chinese BERT learn about syntactic knowledge?

Pre-trained language models such as Bidirectional Encoder Representations from Transformers (BERT) have been applied to a wide range of natural language processing (NLP) tasks and obtained significantly positive results. A growing body of research has investigated the reason why BERT is so efficient...

Full description

Bibliographic Details
Main Authors: Jianyu Zheng, Ying Liu
Format: Article
Language:English
Published: PeerJ Inc. 2023-07-01
Series:PeerJ Computer Science
Subjects:
Online Access:https://peerj.com/articles/cs-1478.pdf