TY - JOUR
T1 - Diversifying Knowledge Enhancement of Biomedical Language Models Using Adapter Modules and Knowledge Graphs
AU - Vladika, Juraj
AU - Fichtl, Alexander
AU - Matthes, Florian
N1 - Publisher Copyright:
© 2024 by SCITEPRESS – Science and Technology Publications, Lda.
PY - 2024
Y1 - 2024
N2 - Recent advances in natural language processing (NLP) owe their success to pre-training language models on large amounts of unstructured data. Still, there is an increasing effort to combine the unstructured nature of LMs with structured knowledge and reasoning. Particularly in the rapidly evolving field of biomedical NLP, knowledge-enhanced language models (KELMs) have emerged as promising tools to bridge the gap between large language models and domain-specific knowledge, considering the available biomedical knowledge graphs (KGs) curated by experts over the decades. In this paper, we develop an approach that uses lightweight adapter modules to inject structured biomedical knowledge into pre-trained language models (PLMs). We use two large KGs, the biomedical knowledge system UMLS and the novel biochemical ontology OntoChem, with two prominent biomedical PLMs, BERT and BioLinkBERT. The approach includes partitioning knowledge graphs into smaller subgraphs, fine-tuning adapter modules for each subgraph, and combining the knowledge in a fusion layer. We test the performance on three downstream tasks: document classification, question answering, and natural language inference. We show that our methodology leads to performance improvements in several instances while keeping requirements in computing power low. Finally, we provide a detailed interpretation of the results and report valuable insights for future work.
AB - Recent advances in natural language processing (NLP) owe their success to pre-training language models on large amounts of unstructured data. Still, there is an increasing effort to combine the unstructured nature of LMs with structured knowledge and reasoning. Particularly in the rapidly evolving field of biomedical NLP, knowledge-enhanced language models (KELMs) have emerged as promising tools to bridge the gap between large language models and domain-specific knowledge, considering the available biomedical knowledge graphs (KGs) curated by experts over the decades. In this paper, we develop an approach that uses lightweight adapter modules to inject structured biomedical knowledge into pre-trained language models (PLMs). We use two large KGs, the biomedical knowledge system UMLS and the novel biochemical ontology OntoChem, with two prominent biomedical PLMs, BERT and BioLinkBERT. The approach includes partitioning knowledge graphs into smaller subgraphs, fine-tuning adapter modules for each subgraph, and combining the knowledge in a fusion layer. We test the performance on three downstream tasks: document classification, question answering, and natural language inference. We show that our methodology leads to performance improvements in several instances while keeping requirements in computing power low. Finally, we provide a detailed interpretation of the results and report valuable insights for future work.
KW - Adapters
KW - Biomedical NLP
KW - Biomedicine
KW - Domain Knowledge
KW - Knowledge Enhancement
KW - Knowledge Graphs
KW - Natural Language Processing (NLP)
KW - Pre-Trained Language Models
UR - http://www.scopus.com/inward/record.url?scp=85190687933&partnerID=8YFLogxK
U2 - 10.5220/0012395200003636
DO - 10.5220/0012395200003636
M3 - Conference article
AN - SCOPUS:85190687933
SN - 2184-3589
VL - 2
SP - 376
EP - 387
JO - International Conference on Agents and Artificial Intelligence
JF - International Conference on Agents and Artificial Intelligence
T2 - 16th International Conference on Agents and Artificial Intelligence, ICAART 2024
Y2 - 24 February 2024 through 26 February 2024
ER -