Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining
Biomedical terms extracted using Word2vec, the most popular word embedding model in recent years, serve as the foundation for various natural language processing (NLP) applications, such as biomedical information retrieval, relation extraction, and recommendation systems. The objective of this study...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2021-12-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/12/1/154 |
_version_ | 1827668761827606528 |
---|---|
author | Ziheng Zhang Feng Han Hongjian Zhang Tomohiro Aoki Katsuhiko Ogasawara |
author_facet | Ziheng Zhang Feng Han Hongjian Zhang Tomohiro Aoki Katsuhiko Ogasawara |
author_sort | Ziheng Zhang |
collection | DOAJ |
description | Biomedical terms extracted using Word2vec, the most popular word embedding model in recent years, serve as the foundation for various natural language processing (NLP) applications, such as biomedical information retrieval, relation extraction, and recommendation systems. The objective of this study is to examine how changes in the ratio of the biomedical domain to general domain data in the corpus affect the extraction of similar biomedical terms using Word2vec. We downloaded abstracts of 214,892 articles from PubMed Central (PMC) and the 3.9 GB Billion Word (BW) benchmark corpus from the computer science community. The datasets were preprocessed and grouped into 11 corpora based on the ratio of BW to PMC, ranging from 0:10 to 10:0, and then Word2vec models were trained on these corpora. The cosine similarities between the biomedical terms obtained from the Word2vec models were then compared in each model. The results indicated that the models trained with both BW and PMC data outperformed the model trained only with medical data. The similarity between the biomedical terms extracted by the Word2vec model increased when the ratio of the biomedical domain to general domain data was 3:7 to 5:5. This study allows NLP researchers to apply Word2vec based on more information and increase the similarity of extracted biomedical terms to improve their effectiveness in NLP applications, such as biomedical information extraction. |
first_indexed | 2024-03-10T03:50:10Z |
format | Article |
id | doaj.art-479941e7dafe4660a29bb05dd3ce0084 |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-10T03:50:10Z |
publishDate | 2021-12-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-479941e7dafe4660a29bb05dd3ce00842023-11-23T11:08:34ZengMDPI AGApplied Sciences2076-34172021-12-0112115410.3390/app12010154Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature MiningZiheng Zhang0Feng Han1Hongjian Zhang2Tomohiro Aoki3Katsuhiko Ogasawara4Graduate School of Health Science, Hokkaido University, Sapporo 060-0808, JapanGraduate School of Health Science, Hokkaido University, Sapporo 060-0808, JapanGraduate School of Health Science, Hokkaido University, Sapporo 060-0808, JapanGraduate School of Health Science, Hokkaido University, Sapporo 060-0808, JapanGraduate School of Health Science, Hokkaido University, Sapporo 060-0808, JapanBiomedical terms extracted using Word2vec, the most popular word embedding model in recent years, serve as the foundation for various natural language processing (NLP) applications, such as biomedical information retrieval, relation extraction, and recommendation systems. The objective of this study is to examine how changes in the ratio of the biomedical domain to general domain data in the corpus affect the extraction of similar biomedical terms using Word2vec. We downloaded abstracts of 214,892 articles from PubMed Central (PMC) and the 3.9 GB Billion Word (BW) benchmark corpus from the computer science community. The datasets were preprocessed and grouped into 11 corpora based on the ratio of BW to PMC, ranging from 0:10 to 10:0, and then Word2vec models were trained on these corpora. The cosine similarities between the biomedical terms obtained from the Word2vec models were then compared in each model. The results indicated that the models trained with both BW and PMC data outperformed the model trained only with medical data. The similarity between the biomedical terms extracted by the Word2vec model increased when the ratio of the biomedical domain to general domain data was 3:7 to 5:5. This study allows NLP researchers to apply Word2vec based on more information and increase the similarity of extracted biomedical terms to improve their effectiveness in NLP applications, such as biomedical information extraction.https://www.mdpi.com/2076-3417/12/1/154biomedical literature mining (BLM)natural language processing (NLP)Word2vec |
spellingShingle | Ziheng Zhang Feng Han Hongjian Zhang Tomohiro Aoki Katsuhiko Ogasawara Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining Applied Sciences biomedical literature mining (BLM) natural language processing (NLP) Word2vec |
title | Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining |
title_full | Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining |
title_fullStr | Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining |
title_full_unstemmed | Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining |
title_short | Examining the Effect of the Ratio of Biomedical Domain to General Domain Data in Corpus in Biomedical Literature Mining |
title_sort | examining the effect of the ratio of biomedical domain to general domain data in corpus in biomedical literature mining |
topic | biomedical literature mining (BLM) natural language processing (NLP) Word2vec |
url | https://www.mdpi.com/2076-3417/12/1/154 |
work_keys_str_mv | AT zihengzhang examiningtheeffectoftheratioofbiomedicaldomaintogeneraldomaindataincorpusinbiomedicalliteraturemining AT fenghan examiningtheeffectoftheratioofbiomedicaldomaintogeneraldomaindataincorpusinbiomedicalliteraturemining AT hongjianzhang examiningtheeffectoftheratioofbiomedicaldomaintogeneraldomaindataincorpusinbiomedicalliteraturemining AT tomohiroaoki examiningtheeffectoftheratioofbiomedicaldomaintogeneraldomaindataincorpusinbiomedicalliteraturemining AT katsuhikoogasawara examiningtheeffectoftheratioofbiomedicaldomaintogeneraldomaindataincorpusinbiomedicalliteraturemining |