MaterialBERT for natural language processing of materials science texts
A BERT (Bidirectional Encoder Representations from Transformers) model, which we named “MaterialBERT”, has been generated using scientific papers in wide area of material science as a corpus. A new vocabulary list for tokenizer was generated using material science corpus. Two BERT models with differ...
Main Authors: | Michiko Yoshitake, Fumitaka Sato, Hiroyuki Kawano, Hiroshi Teraoka |
---|---|
Format: | Article |
Language: | English |
Published: |
Taylor & Francis Group
2022-12-01
|
Series: | Science and Technology of Advanced Materials: Methods |
Subjects: | |
Online Access: | http://dx.doi.org/10.1080/27660400.2022.2124831 |
Similar Items
-
GeoBERT: Pre-Training Geospatial Representation Learning on Point-of-Interest
by: Yunfan Gao, et al.
Published: (2022-12-01) -
Learned Text Representation for Amharic Information Retrieval and Natural Language Processing
by: Tilahun Yeshambel, et al.
Published: (2023-03-01) -
Pre-Training MLM Using Bert for the Albanian Language
by: Kryeziu Labehat, et al.
Published: (2023-06-01) -
Towards Transfer Learning Techniques—BERT, DistilBERT, BERTimbau, and DistilBERTimbau for Automatic Text Classification from Different Languages: A Case Study
by: Rafael Silva Barbon, et al.
Published: (2022-10-01) -
Sentiment Analysis of Text Reviews Using Lexicon-Enhanced Bert Embedding (LeBERT) Model with Convolutional Neural Network
by: James Mutinda, et al.
Published: (2023-01-01)