Scientific question answering remains a significant challenge for the current generation of large language models (LLMs) due to the requirement of engaging with highly specialised concepts. A promising solution is to integrate LLMs with knowledge graphs of research concepts, ensuring that responses are grounded in structured, verifiable information. One effective approach involves using LLMs to translate questions posed in natural language into SPARQL queries, enabling the retrieval of relevant data. In this paper, we analyse the performance of several LLMs on this task using two scientific question-answering benchmarks: SciQA and DBLP-QuAD. We explore both few-shot learning and fine-tuning strategies, investigate error patterns across different models, and propose directions for future research.

Meloni, A., Recupero, D., Osborne, F., Salatino, A., Motta, E., Vahadati, S., et al. (2025). Assessing Large Language Models for SPARQL Query Generation in Scientific Question Answering. In Proceedings of the Special Session on Harmonising Generative AI and Semantic Web Technologies (HGAIS 2024) co-located with the 23rd International Semantic Web Conference (ISWC 2024) (pp.1-7). CEUR-WS.

Assessing Large Language Models for SPARQL Query Generation in Scientific Question Answering

Osborne F.;
2025

Abstract

Scientific question answering remains a significant challenge for the current generation of large language models (LLMs) due to the requirement of engaging with highly specialised concepts. A promising solution is to integrate LLMs with knowledge graphs of research concepts, ensuring that responses are grounded in structured, verifiable information. One effective approach involves using LLMs to translate questions posed in natural language into SPARQL queries, enabling the retrieval of relevant data. In this paper, we analyse the performance of several LLMs on this task using two scientific question-answering benchmarks: SciQA and DBLP-QuAD. We explore both few-shot learning and fine-tuning strategies, investigate error patterns across different models, and propose directions for future research.
paper
Knowledge Graphs; Large Language Models; Machine Translation; SPARQL;
English
2024 Harmonising Generative AI and Semantic Web Technologies, HGAIS 2024 - November 13, 2024
2024
Alharbi, R; de Berardinis, J; Groth, P; Meroño Peñuela, A; Simperl , E; Tamma, V
Proceedings of the Special Session on Harmonising Generative AI and Semantic Web Technologies (HGAIS 2024) co-located with the 23rd International Semantic Web Conference (ISWC 2024)
2025
3953
1
7
https://ceur-ws.org/Vol-3953/
open
Meloni, A., Recupero, D., Osborne, F., Salatino, A., Motta, E., Vahadati, S., et al. (2025). Assessing Large Language Models for SPARQL Query Generation in Scientific Question Answering. In Proceedings of the Special Session on Harmonising Generative AI and Semantic Web Technologies (HGAIS 2024) co-located with the 23rd International Semantic Web Conference (ISWC 2024) (pp.1-7). CEUR-WS.
File in questo prodotto:
File Dimensione Formato  
Meloni-2025-HGAIS-VoR.pdf

accesso aperto

Descrizione: This volume and its papers are published under the Creative Commons License Attribution 4.0 International (CC BY 4.0).
Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Creative Commons
Dimensione 209.26 kB
Formato Adobe PDF
209.26 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/553722
Citazioni
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
Social impact