View a PDF of the paper titled Investigating LLM Capabilities on Long Context Comprehension for Medical Question Answering, by Feras AlMannaa and 3 other authors
Abstract:This study is the first to investigate LLM comprehension capabilities over long-context (LC), clinically relevant medical Question Answering (QA) beyond MCQA. Our comprehensive approach considers a range of settings based on content inclusion of varying size and relevance, LLM models of different capabilities and a variety of datasets across task formulations. We reveal insights on model size effects and their limitations, underlying memorization issues and the benefits of reasoning models, while demonstrating the value and challenges of leveraging the full long patient’s context. Importantly, we examine the effect of Retrieval Augmented Generation (RAG) on medical LC comprehension, showcasing best settings in single versus multi-document QA datasets. We shed light into some of the evaluation aspects using a multi-faceted approach uncovering common metric challenges. Our quantitative analysis reveals challenging cases where RAG excels while still showing limitations in cases requiring temporal reasoning.
Submission history
From: Feras AlMannaa [view email]
[v1]
Tue, 21 Oct 2025 14:50:24 UTC (2,482 KB)
[v2]
Wed, 14 Jan 2026 21:40:27 UTC (2,585 KB)
Source link
#Investigating #LLM #Capabilities #Long #Context #Comprehension #Medical #Question #Answering
















![[2510.18691] Investigating LLM Capabilities on Long Context Comprehension for Medical Question Answering [2510.18691] Investigating LLM Capabilities on Long Context Comprehension for Medical Question Answering](https://i0.wp.com/arxiv.org/static/browse/0.3.4/images/arxiv-logo-fb.png?w=750&resize=750,375&ssl=1)








