Enhancing SPARQL Query Performance With Recurrent Neural Networks

DBpedia is one of the most resourceful link databases today, and users need to use query syntax (e.g., SPARQL) to access information in DBpedia databases. However, not all users know SPARQL, so a natural language query system can be used to translate the user’s query into the correspondin...

Full description

Bibliographic Details
Main Authors: Yi-Hui Chen, Eric Jui-Lin Lu, Jin-De Lin
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10230082/
_version_ 1797692169380691968
author Yi-Hui Chen
Eric Jui-Lin Lu
Jin-De Lin
author_facet Yi-Hui Chen
Eric Jui-Lin Lu
Jin-De Lin
author_sort Yi-Hui Chen
collection DOAJ
description DBpedia is one of the most resourceful link databases today, and users need to use query syntax (e.g., SPARQL) to access information in DBpedia databases. However, not all users know SPARQL, so a natural language query system can be used to translate the user’s query into the corresponding syntax. Generating query syntax through the query system is both time-consuming and expensive. To improve the efficiency of query syntax generation from user questions, the multi-label template approach, specifically Light-QAwizard, is utilized. Light-QAwizard transforms the problem into one or more single-label classifications using multi-label learning template approaches. By implementing Light-QAwizard, query costs can be reduced by 50%, but it introduces a new label during the transformation process leading to sample imbalance, compromised accuracy, and limited scalability. To overcome these limitations, this paper employs two multi-label learning methods, Binary Relevance (BR) and Classifier Chains (CC), for question transformation. By employing Recurrent Neural Networks (RNNs) as a multi-label classifier for generating RDF (Resource Description Framework) triples, all the labels are predicted to align with the query intentions. To better account for the relationship between RDF triples, BR is integrated into an ensemble learning approach to result in the Ensemble BR. Experimental results demonstrate that our proposed method outperforms previous research in terms of improving query accuracy. The favorable experiments substantiate that the Ensemble BR or CC model demonstrates competitiveness by integrating label relationships into the trained model.
first_indexed 2024-03-12T02:23:44Z
format Article
id doaj.art-0c4858f2f62a4c639462941cb8889a45
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-03-12T02:23:44Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-0c4858f2f62a4c639462941cb8889a452023-09-05T23:00:48ZengIEEEIEEE Access2169-35362023-01-0111922099222410.1109/ACCESS.2023.330869110230082Enhancing SPARQL Query Performance With Recurrent Neural NetworksYi-Hui Chen0https://orcid.org/0000-0002-9932-0594Eric Jui-Lin Lu1https://orcid.org/0000-0001-7953-5486Jin-De Lin2Department of Information Management, Chang Gung University, Taoyuan, TaiwanDepartment of Management Information Systems, National Chung Hsing University, Taichung City, TaiwanDepartment of Management Information Systems, National Chung Hsing University, Taichung City, TaiwanDBpedia is one of the most resourceful link databases today, and users need to use query syntax (e.g., SPARQL) to access information in DBpedia databases. However, not all users know SPARQL, so a natural language query system can be used to translate the user’s query into the corresponding syntax. Generating query syntax through the query system is both time-consuming and expensive. To improve the efficiency of query syntax generation from user questions, the multi-label template approach, specifically Light-QAwizard, is utilized. Light-QAwizard transforms the problem into one or more single-label classifications using multi-label learning template approaches. By implementing Light-QAwizard, query costs can be reduced by 50%, but it introduces a new label during the transformation process leading to sample imbalance, compromised accuracy, and limited scalability. To overcome these limitations, this paper employs two multi-label learning methods, Binary Relevance (BR) and Classifier Chains (CC), for question transformation. By employing Recurrent Neural Networks (RNNs) as a multi-label classifier for generating RDF (Resource Description Framework) triples, all the labels are predicted to align with the query intentions. To better account for the relationship between RDF triples, BR is integrated into an ensemble learning approach to result in the Ensemble BR. Experimental results demonstrate that our proposed method outperforms previous research in terms of improving query accuracy. The favorable experiments substantiate that the Ensemble BR or CC model demonstrates competitiveness by integrating label relationships into the trained model.https://ieeexplore.ieee.org/document/10230082/DBPediaSPARQLquestion answering systemsmulti-label classifierrecurrent neural network (RNN)binary relevance (BR)
spellingShingle Yi-Hui Chen
Eric Jui-Lin Lu
Jin-De Lin
Enhancing SPARQL Query Performance With Recurrent Neural Networks
IEEE Access
DBPedia
SPARQL
question answering systems
multi-label classifier
recurrent neural network (RNN)
binary relevance (BR)
title Enhancing SPARQL Query Performance With Recurrent Neural Networks
title_full Enhancing SPARQL Query Performance With Recurrent Neural Networks
title_fullStr Enhancing SPARQL Query Performance With Recurrent Neural Networks
title_full_unstemmed Enhancing SPARQL Query Performance With Recurrent Neural Networks
title_short Enhancing SPARQL Query Performance With Recurrent Neural Networks
title_sort enhancing sparql query performance with recurrent neural networks
topic DBPedia
SPARQL
question answering systems
multi-label classifier
recurrent neural network (RNN)
binary relevance (BR)
url https://ieeexplore.ieee.org/document/10230082/
work_keys_str_mv AT yihuichen enhancingsparqlqueryperformancewithrecurrentneuralnetworks
AT ericjuilinlu enhancingsparqlqueryperformancewithrecurrentneuralnetworks
AT jindelin enhancingsparqlqueryperformancewithrecurrentneuralnetworks