Transformer-based approach to variable typing

The upsurge of multifarious endeavors across scientific fields propelled Big Data in the scientific domain. Despite the advancements in management systems, researchers find that mathematical knowledge remains one of the most challenging to manage due to the latter's inherent heterogeneity. One...

Full description

Bibliographic Details
Main Authors: Charles Arthel Rey, Jose Lorenzo Danguilan, Karl Patrick Mendoza, Miguel Francisco Remolona
Format: Article
Language:English
Published: Elsevier 2023-10-01
Series:Heliyon
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2405844023077137
_version_ 1797646523957248000
author Charles Arthel Rey
Jose Lorenzo Danguilan
Karl Patrick Mendoza
Miguel Francisco Remolona
author_facet Charles Arthel Rey
Jose Lorenzo Danguilan
Karl Patrick Mendoza
Miguel Francisco Remolona
author_sort Charles Arthel Rey
collection DOAJ
description The upsurge of multifarious endeavors across scientific fields propelled Big Data in the scientific domain. Despite the advancements in management systems, researchers find that mathematical knowledge remains one of the most challenging to manage due to the latter's inherent heterogeneity. One novel recourse being explored is variable typing where current works remain preliminary and, thus, provide a wide room for contribution. In this study, a primordial attempt to implement the end-to-end Entity Recognition (ER) and Relation Extraction (RE) approach to variable typing was made using the BERT (Bidirectional Encoder Representations from Transformers) model. A micro-dataset was developed for this process. According to our findings, the ER model and RE model, respectively, have Precision of 0.8142 and 0.4919, Recall of 0.7816 and 0.6030, and F1-Scores of 0.7975 and 0.5418. Despite the limited dataset, the models performed at par with values in the literature. This work also discusses the factors affecting this BERT-based approach, giving rise to suggestions for future implementations.
first_indexed 2024-03-11T15:03:55Z
format Article
id doaj.art-7ed1d61911724571be0b87892b07456c
institution Directory Open Access Journal
issn 2405-8440
language English
last_indexed 2024-03-11T15:03:55Z
publishDate 2023-10-01
publisher Elsevier
record_format Article
series Heliyon
spelling doaj.art-7ed1d61911724571be0b87892b07456c2023-10-30T06:06:23ZengElsevierHeliyon2405-84402023-10-01910e20505Transformer-based approach to variable typingCharles Arthel Rey0Jose Lorenzo Danguilan1Karl Patrick Mendoza2Miguel Francisco Remolona3Chemical Engineering Intelligence Learning Laboratory, Department of Chemical Engineering, University of the Philippines Diliman, Quezon City, 1101 PhilippinesChemical Engineering Intelligence Learning Laboratory, Department of Chemical Engineering, University of the Philippines Diliman, Quezon City, 1101 PhilippinesChemical Engineering Intelligence Learning Laboratory, Department of Chemical Engineering, University of the Philippines Diliman, Quezon City, 1101 PhilippinesCorresponding author.; Chemical Engineering Intelligence Learning Laboratory, Department of Chemical Engineering, University of the Philippines Diliman, Quezon City, 1101 PhilippinesThe upsurge of multifarious endeavors across scientific fields propelled Big Data in the scientific domain. Despite the advancements in management systems, researchers find that mathematical knowledge remains one of the most challenging to manage due to the latter's inherent heterogeneity. One novel recourse being explored is variable typing where current works remain preliminary and, thus, provide a wide room for contribution. In this study, a primordial attempt to implement the end-to-end Entity Recognition (ER) and Relation Extraction (RE) approach to variable typing was made using the BERT (Bidirectional Encoder Representations from Transformers) model. A micro-dataset was developed for this process. According to our findings, the ER model and RE model, respectively, have Precision of 0.8142 and 0.4919, Recall of 0.7816 and 0.6030, and F1-Scores of 0.7975 and 0.5418. Despite the limited dataset, the models performed at par with values in the literature. This work also discusses the factors affecting this BERT-based approach, giving rise to suggestions for future implementations.http://www.sciencedirect.com/science/article/pii/S2405844023077137Natural language processingTransformersEntity recognitionRelation extractionVariable typingMachine learning
spellingShingle Charles Arthel Rey
Jose Lorenzo Danguilan
Karl Patrick Mendoza
Miguel Francisco Remolona
Transformer-based approach to variable typing
Heliyon
Natural language processing
Transformers
Entity recognition
Relation extraction
Variable typing
Machine learning
title Transformer-based approach to variable typing
title_full Transformer-based approach to variable typing
title_fullStr Transformer-based approach to variable typing
title_full_unstemmed Transformer-based approach to variable typing
title_short Transformer-based approach to variable typing
title_sort transformer based approach to variable typing
topic Natural language processing
Transformers
Entity recognition
Relation extraction
Variable typing
Machine learning
url http://www.sciencedirect.com/science/article/pii/S2405844023077137
work_keys_str_mv AT charlesarthelrey transformerbasedapproachtovariabletyping
AT joselorenzodanguilan transformerbasedapproachtovariabletyping
AT karlpatrickmendoza transformerbasedapproachtovariabletyping
AT miguelfranciscoremolona transformerbasedapproachtovariabletyping