Core Concepts
MEG is a novel approach that enhances the accuracy of large language models (LLMs) in medical question answering by efficiently integrating medical knowledge graph embeddings.
Stats
MEG achieves an average of +10.2% accuracy over the Mistral-Instruct baseline, and +6.7% over specialized models like BioMistral on four medical question answering datasets.
MEG-MISTRAL1 achieves 54.6% accuracy on MedQA, 74.6% on PubMedQA, 56.4% on MedMCQA, and 60.3% on MMLU-Medical.
MEG-LLAMA achieves 66.0% accuracy on MedQA, 78.0% on PubMedQA, 60.6% on MedMCQA, and 74.9% on MMLU-Medical.
The mapping network in MEG has 1.22M parameters.
Phase I of training takes 4 hours on 4 NVIDIA A10G GPUs.
Quotes
"LLMs, which underpin most contemporary question answering systems, struggle to induce how concepts relate in specialized domains such as medicine."
"MEG, a parameter-efficient approach for medical knowledge-augmented LLMs."
"MEG uses a lightweight mapping network to integrate graph embeddings into the LLM, enabling it to leverage external knowledge in a cost-effective way."
"We evaluate our method on four popular medical multiple-choice datasets and show that LLMs greatly benefit from the factual grounding provided by knowledge graph embeddings."