On Constructing Biomedical Text-to-Graph Systems with Large Language Models
Knowledge graphs and ontologies represent symbolic and factual information that can offer structured and interpretable knowledge. Extracting and manipulating this type of information is a crucial step in complex processes such as human reasoning. While Large Language Models (LLMs) are known to be useful for extracting and enriching knowledge graphs and ontologies, previous work has largely focused on comparing architecture-specific models (e.g. encoder-decoder only) across benchmarks from similar domains. In this work, we provide a large-scale comparison of the performance of certain LLM features (e.g. model architecture and size) and task learning methods (fine-tuning vs. in-context learning (iCL)) on text-to-graph benchmarks in the biomedical domain. Our experiment suggests that, while a simple truncation-based heuristic can notably boost the performance of decoder-only models used with iCL, small fine-tuned encoder-decoder models produce the most stable and strong performance. Moreover, we found that a massive out-of-domain text-graph pre-training has a positive impact on fine-tuned models, while we observed only a marginal impact of pre-training and size for decoder-only iCL models.
BERTOLINI Lorenzo;
HULSMAN Roel;
CONSOLI Sergio;
PUERTAS GALLARDO Antonio;
CERESA Mario;
2024-09-03
Elsevier
JRC137500
1613-0073 (online),
https://ceur-ws.org/Vol-3747/text2kg_paper10.pdf,
https://publications.jrc.ec.europa.eu/repository/handle/JRC137500,
Additional supporting files
| File name | Description | File type | |