← Back to papers

Paper deep dive

Grounded Multimodal Retrieval-Augmented Drafting of Radiology Impressions Using Case-Based Similarity Search

Himadri Samanta

Year: 2026Venue: arXiv preprintArea: q-bio.QMType: PreprintEmbeddings: 23

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 97%

Last extracted: 3/23/2026, 12:02:49 PM

Summary

This paper presents a grounded multimodal retrieval-augmented generation (RAG) system for drafting radiology impressions. By combining CLIP-based image and text embeddings with FAISS-indexed similarity search, the system retrieves relevant historical cases to ground the generation of new reports. The approach includes citation verification and confidence-based refusal mechanisms to improve factual consistency and clinical reliability, demonstrating superior retrieval performance compared to image-only methods.

Entities (5)

CLIP · model-architecture · 100%FAISS · software-library · 100%MIMIC-CXR · dataset · 100%RAG · methodology · 100%FastAPI · software-framework · 95%

Relation Signals (4)

FAISS enables Nearest-neighbor retrieval

confidence 100% · A fusion similarity framework was implemented using FAISS indexing for scalable nearest-neighbor retrieval.

CLIP generatesembeddingsfor Chest X-ray images

confidence 100% · Image embeddings were generated using CLIP encoders

MIMIC-CXR usedfor Multimodal retrieval database

confidence 100% · A curated subset of the MIMIC-CXR dataset was used to construct a multimodal retrieval database.

RAG drafts Radiology impressions

confidence 95% · we propose a multimodal retrieval-augmented generation (RAG) system for grounded drafting of chest radiograph impressions.

Cypher Suggestions (2)

Find all datasets used in the research · confidence 90% · unvalidated

MATCH (e:Entity {entity_type: 'Dataset'}) RETURN e.name

Map the relationship between methodologies and their components · confidence 85% · unvalidated

MATCH (s)-[r]->(o) WHERE r.relation IN ['USED_FOR', 'ENABLES', 'GENERATES_EMBEDDINGS_FOR'] RETURN s.name, r.relation, o.name

Abstract

Abstract:Automated radiology report generation has gained increasing attention with the rise of deep learning and large language models. However, fully generative approaches often suffer from hallucinations and lack clinical grounding, limiting their reliability in real-world workflows. In this study, we propose a multimodal retrieval-augmented generation (RAG) system for grounded drafting of chest radiograph impressions. The system combines contrastive image-text embeddings, case-based similarity retrieval, and citation-constrained draft generation to ensure factual alignment with historical radiology reports. A curated subset of the MIMIC-CXR dataset was used to construct a multimodal retrieval database. Image embeddings were generated using CLIP encoders, while textual embeddings were derived from structured impression sections. A fusion similarity framework was implemented using FAISS indexing for scalable nearest-neighbor retrieval. Retrieved cases were used to construct grounded prompts for draft impression generation, with safety mechanisms enforcing citation coverage and confidence-based refusal. Experimental results demonstrate that multimodal fusion significantly improves retrieval performance compared to image-only retrieval, achieving Recall@5 above 0.95 on clinically relevant findings. The grounded drafting pipeline produces interpretable outputs with explicit citation traceability, enabling improved trustworthiness compared to conventional generative approaches. This work highlights the potential of retrieval-augmented multimodal systems for reliable clinical decision support and radiology workflow augmentation

Tags

ai-safety (imported, 100%)preprint (suggested, 88%)q-bioqm (suggested, 92%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Full Text

22,595 characters extracted from source content.

Expand or collapse full text

Grounded Multimodal Retrieval-Augmented Drafting of Radiology Impressions Using Case-Based Similarity Search Himadri Sekhar Samanta Independent AI Researcher, Austin, Texas, USA Abstract Automated radiology report generation has gained increasing attention with the rise of deep learning and large language models. However, fully generative approaches often suffer from hallucinations and lack clinical grounding, limiting their reliability in real-world workflows. In this study, we propose a multimodal retrieval-augmented generation (RAG) system for grounded drafting of chest radiograph impressions. The system combines contrastive image-text embeddings, case-based similarity retrieval, and citation-constrained draft generation to ensure factual alignment with historical radiology reports. A curated subset of the MIMIC-CXR dataset was used to construct a multimodal retrieval database. Image embeddings were generated using CLIP encoders, while textual embeddings were derived from structured impression sections. A fusion similarity framework was implemented using FAISS indexing for scalable nearest-neighbor retrieval. Retrieved cases were used to construct grounded prompts for draft impression generation, with safety mechanisms enforcing citation coverage and confidence-based refusal. Experimental results demonstrate that multimodal fusion significantly improves retrieval performance compared to image-only retrieval, achieving Recall@5 above 0.95 on clinically relevant findings. The grounded drafting pipeline produces interpretable outputs with explicit citation traceability, enabling improved trustworthiness compared to conventional generative approaches. This work highlights the potential of retrieval-augmented multimodal systems for reliable clinical decision support and radiology workflow augmentation. keywords: Radiology report generation , multimodal retrieval , retrieval-augmented generation , chest X-ray , medical imaging , explainable AI †journal: Journal of Biomedical Informatics 1 Introduction Artificial intelligence has shown promise in assisting radiologists with image interpretation and report drafting. Deep learning has substantially improved performance in medical image analysis tasks [13, 14], including thoracic imaging, where chest radiographs are among the most widely performed diagnostic studies [17, 7]. More recently, large language models and multimodal systems have enabled automated clinical text generation, but they remain vulnerable to hallucinations, unsupported claims, and poor calibration in safety-critical settings [15, 11]. These limitations are especially problematic in radiology, where factual consistency and traceability are essential. Retrieval-augmented generation (RAG) offers an alternative by constraining generation to retrieved evidence [12, 5]. Instead of producing free-form text solely from model priors, RAG systems retrieve semantically similar historical examples and condition generation on verifiable evidence. Such case-based reasoning aligns closely with radiology practice, where prior studies and similar historical cases frequently inform interpretation. In radiology, both the chest X-ray image and the report impression are informative: the image captures anatomical and pathological appearance, while the report impression encodes concise clinical interpretation. In this work, we present a grounded multimodal retrieval-augmented system for chest radiograph impression drafting. The proposed framework combines image and report embeddings, performs late fusion for similarity retrieval, and indexes the resulting vectors with FAISS for efficient nearest-neighbor search [9]. Retrieved cases are then used to generate grounded draft impressions with explicit case citations. To improve safety, the system applies a confidence threshold and refuses generation when similarity is insufficient. We additionally deploy the pipeline as a FastAPI service and package it for reproducible Docker-based inference. The main contributions of this work are as follows: 1. We construct a clean multimodal chest X-ray dataset subset from MIMIC-CXR with aligned image paths, study identifiers, and extracted impression text suitable for retrieval and drafting experiments [8]. 2. We demonstrate that multimodal fusion of image and text embeddings markedly improves retrieval performance over image-only retrieval. 3. We introduce a grounded drafting mechanism with explicit case citations and confidence-based refusal to reduce unsupported output generation. 4. We provide an end-to-end deployable implementation, including a REST API and containerized execution environment, illustrating practical translation beyond notebook experimentation. 2 Related Work Automated radiology report generation has been studied using encoder–decoder models, transformer-based image captioning systems, and multimodal vision-language architectures. While these approaches have shown promising fluency, they often struggle with factual consistency and hallucination. Chest X-ray modeling has particularly benefited from large public datasets and weakly supervised labeling pipelines, such as CheXpert [7]. Contrastive language-image pretraining further demonstrated that aligned multimodal embeddings can support strong cross-modal representations for retrieval and transfer learning [16]. Medical-domain adaptations, including BioViL and XrayBERT, highlight the growing role of clinically aligned vision-language pretraining [3, 1]. In parallel, retrieval-based and case-based reasoning methods have been explored in radiology because they naturally align with clinical practice and can improve interpretability. Efficient similarity search frameworks such as FAISS support practical deployment of large embedding databases [9]. More recently, RAG pipelines have been proposed as a way to improve factuality and evidence grounding in language generation [12, 5, 18, 2]. However, relatively few systems combine radiology images and report text within a unified retrieval architecture that also includes citation verification and confidence-based refusal. Our work addresses this gap by integrating multimodal retrieval, grounded drafting, safety gating, and deployable serving in a single workflow. Explainability and deployment realism are also central concerns in clinical AI [6, 4, 19]. Prior work has emphasized that successful healthcare AI must not only achieve strong offline metrics but also provide interpretable outputs, calibrated uncertainty, and operational pathways to translation [11, 21]. Our design choices are motivated by these principles. 3 Dataset We used a curated subset of the MIMIC-CXR dataset and the MIMIC-CXR-JPG release [8]. The development process intentionally avoided downloading the full multi-terabyte dataset. Instead, lightweight metadata tables and the compressed report archive were first downloaded from PhysioNet under credentialed access. A reproducible subset of 2,000 studies was sampled using metadata mappings between subject identifiers, study identifiers, and image identifiers. After expansion to DICOM-level rows, the initial subset contained 3,301 entries. To construct a modeling-ready image dataset, the study manifest was then joined to MIMIC-CXR-JPG metadata in order to retain only studies with available JPG images. This yielded 3,149 rows with corresponding image files. The IMPRESSION section was extracted from each report using rule-based section parsing, as the impression provides the most concise clinically relevant summary for retrieval and draft generation. Rows without a valid impression section were removed, resulting in a final clean multimodal dataset of 2,696 image–impression pairs. Table 1 summarizes the dataset construction process. Table 1: Dataset construction summary. Stage Count Sampled studies 2000 DICOM rows in initial manifest 3301 Rows with available JPG images 3149 Rows with valid impression text 2696 Final clean multimodal dataset 2696 Positive-label evaluation subset 839 4 Methods 4.1 Image representation learning Each chest X-ray image was encoded using the CLIP ViT-B/32 image encoder [16]. Images were loaded as RGB inputs, preprocessed using the CLIP processor, and converted into 512-dimensional embeddings. All image embeddings were L2-normalized to facilitate cosine similarity retrieval through inner-product search. 4.2 Text representation learning For each study, the extracted report impression was encoded using the CLIP text encoder. Texts were tokenized with padding and truncation, embedded into the same 512-dimensional space, and L2-normalized. Because both encoders are aligned through contrastive pretraining, the image and text vectors are directly comparable. 4.3 Multimodal fusion Late fusion was used to combine image and text embeddings: f​u​s​i​o​n=α​i​m​a​g​e+(1−α)​t​e​x​t,e_fusion= _image+(1-α)e_text, where α∈[0,1]α∈[0,1] controls the contribution of the visual branch. The fused vectors were normalized after combination. We evaluated multiple values of α and selected the setting with the best retrieval performance on the development subset. The best-performing setting was α=0.5α=0.5. Figure 3 summarizes the fusion-weight ablation. 4.4 Retrieval system All normalized fused embeddings were indexed using FAISS with an inner-product similarity index [9]. Because the vectors were normalized, inner product is equivalent to cosine similarity. At inference time, the system computes an embedding for the query image and retrieves the top-K most similar cases. For deployment, we used an image-only query against the image-side index to avoid a modality mismatch between query-time inputs and indexed vectors, while multimodal fusion was retained for offline evaluation and ablation. 4.5 Grounded draft generation For each query, the top-K retrieved cases were used as supporting evidence. Retrieved report impressions were converted into short evidence snippets, and a grounded draft impression was produced either by a lightweight language model or by a deterministic evidence-based summarizer. The deterministic fallback ensures that the draft remains constrained to retrieved evidence and can be used when the language model output is malformed or insufficiently grounded. 4.6 Citation verification Each draft output was required to reference retrieved evidence using explicit case identifiers (e.g., [Case 1], [Case 2]). Citation coverage was computed as the fraction of expected case markers present in the final draft. Missing citations were recorded for error analysis. This provides a simple but effective mechanism to track evidence attribution. 4.7 Confidence-based refusal The similarity score of the top retrieved case was used as a confidence signal. When the top-1 similarity score fell below a predefined threshold, the system refused to generate a report draft and instead returned a structured refusal response. This mechanism prevents low-confidence or out-of-distribution inputs from producing unsupported clinical text. 4.8 Deployment The final system was deployed as a FastAPI-based REST service with two primary endpoints: /health for health checks and /predict for inference. The service returns prediction status, confidence score, latency, generated draft, and retrieved case identifiers. For reproducibility, the system was packaged in Docker with a version-controlled dependency file. 5 Experiments 5.1 Retrieval evaluation We evaluated retrieval quality using Recall@K with K∈1,5,10K∈\1,5,10\. Relevance was defined using CheXpert-derived pathology labels [7]. A query was considered successful if at least one retrieved case among the top-K shared the same positive clinical label as the query. We additionally compared image-only retrieval against multimodal fusion retrieval. 5.2 Fusion weight study To understand the contribution of image and text signals, we performed an alpha sweep over fusion weights and measured Recall@5. The best setting was retained for subsequent experiments. 5.3 Safety evaluation We evaluated the deployed RAG layer using refusal rate, average top-1 retrieval similarity, and average citation coverage. Refusal rate quantifies how often the safety policy suppresses generation; citation coverage measures the extent to which the final report explicitly references retrieved evidence. 6 Results 6.1 Retrieval performance Multimodal fusion substantially improved retrieval performance over image-only retrieval. Table 2 summarizes the main retrieval metrics, while Figures 1, 2, and 3 provide visual summaries. Table 2: Retrieval performance comparison. Method Recall@1 Recall@5 Recall@10 Image-only — 0.633 — Fusion (α=0.5α=0.5) 0.739 0.956 0.981 Best fusion setting — 0.975 — Figure 1: Recall@5 comparison between image-only retrieval and multimodal fusion. Figure 2: Fusion retrieval performance across different values of K. Figure 3: Fusion-weight ablation showing peak retrieval performance near α=0.5α=0.5. Image-only retrieval achieved Recall@5 of 0.633, indicating that visual similarity alone captures some clinically useful signal but misses a substantial fraction of relevant cases. In contrast, multimodal fusion achieved Recall@5 of 0.975 at the best fusion setting and Recall@5 of 0.956 on the positive-case evaluation subset. These results indicate that report semantics provide complementary information beyond image appearance alone. 6.2 Safety and reliability The deployed RAG system demonstrated strong safety metrics. As shown in Table 3 and Figure 4, the refusal rate on the internal evaluation set was 0.000, average top-1 retrieval similarity was 0.980, and average citation coverage was 0.867. These results indicate that most in-distribution chest X-ray inputs retrieve strong matches and that the majority of generated statements remain explicitly grounded in cited evidence. Table 3: Safety and grounding metrics. Metric Value Refusal rate 0.000 Average top-1 retrieval score 0.980 Average citation coverage 0.867 Best fusion weight α 0.5 Figure 4: Safety and grounding metrics: refusal rate, average best retrieval score, and average citation coverage. 6.3 Qualitative example Table 4 presents a representative case illustrating the proposed grounded drafting approach. Table 4: Representative qualitative example of grounded drafting. Component Example Query finding Mild bibasilar atelectatic change on frontal chest radiograph Retrieved Case 1 Bibasilar atelectasis. Otherwise, no acute cardiopulmonary abnormality. Retrieved Case 2 Mild bibasilar atelectasis. No other acute findings. Retrieved Case 3 Mild bibasilar atelectasis. Otherwise, no acute cardiopulmonary process. Generated draft Mild bibasilar atelectasis. [Case 1][Case 2] No acute cardiopulmonary abnormality. [Case 1][Case 3] 6.4 Architecture overview Figure 5 illustrates the complete multimodal RAG pipeline used in this work, from image and text encoding to retrieval, grounded drafting, citation verification, confidence gating, and API output. Figure 5: System architecture for grounded multimodal retrieval-augmented radiology drafting. 7 Error Analysis The main error patterns observed in development fell into three categories. First, image-only retrieval sometimes returned anatomically similar but clinically different cases, especially when subtle findings required report semantics for disambiguation. Second, the lightweight language model occasionally copied evidence markers or repeated retrieved content verbatim rather than producing a concise synthesized impression. Third, some drafts omitted one or more expected citations, reducing citation coverage. These issues motivated the use of a deterministic evidence-based fallback and explicit citation verification. We also verified the behavior of the confidence-based refusal mechanism on out-of-domain images. When a non-chest photograph was submitted, the system returned a low similarity score and correctly refused to generate a draft under the default clinical threshold. Lowering the threshold enabled generation in demo mode, but this setting is not recommended for clinical use. 8 Discussion This study demonstrates that multimodal retrieval is a strong foundation for safer radiology draft generation. The retrieval results suggest that the textual impression signal contributes critical clinical information that image-only representations do not fully capture. The strong Recall@5 and Recall@10 results indicate that the fused embedding space successfully organizes studies according to clinically meaningful similarity. The RAG layer provides an important practical advantage: rather than asking a model to generate a report from scratch, the system constrains drafting to retrieved, previously observed evidence. This substantially improves interpretability. In addition, confidence gating provides a pragmatic safety control by allowing the system to abstain when retrieval evidence is weak. Together, these components align with real clinical requirements for traceability and conservative behavior [6, 11]. The deployed API and Docker packaging are also significant from an applied perspective. Many academic systems stop at offline metrics, whereas our implementation demonstrates a path toward reproducible and portable deployment. This increases the work’s relevance to both academic and industry audiences [4, 19, 21]. 9 Limitations This work has several limitations. First, experiments were conducted on a curated subset rather than the full MIMIC-CXR dataset, which may limit generalization. Second, the report drafting component relied on simple prompt-based or deterministic summarization rather than a radiology-specialized large language model. Third, evaluation emphasized retrieval and evidence-grounding metrics, but did not yet include formal radiologist review of generated outputs. Finally, citation coverage is a useful but imperfect proxy for factual correctness; future work should incorporate stronger faithfulness evaluation, larger-scale validation, and external benchmarking [20, 10]. 10 Conclusion We presented a grounded multimodal retrieval-augmented radiology co-pilot for chest X-ray impression drafting. The proposed framework combines image and report embeddings, uses FAISS for efficient case retrieval, and generates citation-grounded drafts with a confidence-based refusal policy. Experiments on a curated MIMIC-CXR subset show that multimodal fusion greatly improves retrieval performance over image-only baselines and that the deployed system achieves strong retrieval confidence and citation coverage. These results support the use of multimodal retrieval and explicit evidence grounding as promising directions for safer and more interpretable AI assistance in radiology workflows. References [1] B. Boecking, N. Usuyama, S. Bannur, S. Hyland, Z. Liu, et al. (2022) Making the most of text semantics to improve biomedical vision-language processing. EMNLP. Cited by: §2. [2] Q. Chen et al. (2023) Clinical retrieval-augmented generation for evidence-based decision support. NPJ Digital Medicine. Cited by: §2. [3] J. Delbrouck, P. Chambon, G. Gohy, P. Sounack, J. Chaves, et al. (2022) BioViL: a knowledge-enriched vision-language model for medical image understanding and generation. Scientific Reports. Cited by: §2. [4] A. Esteva, A. Robicquet, B. Ramsundar, V. Kuleshov, M. DePristo, et al. (2019) A guide to deep learning in healthcare. Nature Medicine 25, p. 24–29. Cited by: §2, §8. [5] Y. Gao, Y. Xiong, X. Gao, K. Jia, J. Pan, et al. (2024) Retrieval-augmented generation for large language models: a survey. arXiv preprint arXiv:2312.10997. Cited by: §1, §2. [6] A. Holzinger, C. Biemann, C. Pattichis, and D. Kell (2017) What do we need to build explainable ai systems for the medical domain?. arXiv preprint arXiv:1712.09923. Cited by: §2, §8. [7] J. Irvin, P. Rajpurkar, M. Ko, Y. Yu, S. Ciurea-Ilcus, et al. (2019) CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. Proceedings of AAAI. Cited by: §1, §2, §5.1. [8] A. E.W. Johnson, T. J. Pollard, S. J. Berkowitz, N. R. Greenbaum, M. P. Lungren, C. Deng, R. G. Mark, and S. Horng (2019) MIMIC-cxr, a de-identified publicly available database of chest radiographs with free-text reports. Scientific Data 6 (317). Cited by: item 1, §3. [9] J. Johnson, M. Douze, and H. Jegou (2021) Billion-scale similarity search with gpus. IEEE Transactions on Big Data. Cited by: §1, §2, §4.4. [10] A. Karargyris, J. T. Wu, A. Sharma, M. Morris, et al. (2021) RadImageNet: an open radiologic deep learning research dataset for effective transfer learning. Radiology: Artificial Intelligence 3 (5). Cited by: §9. [11] C. Kelly, A. Karthikesalingam, M. Suleyman, G. Corrado, and D. King (2019) Key challenges for delivering clinical impact with artificial intelligence. BMC Medicine 17 (195). Cited by: §1, §2, §8. [12] P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, et al. (2020) Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems. Cited by: §1, §2. [13] G. Litjens, T. Kooi, B. Bejnordi, A. Setio, F. Ciompi, et al. (2017) A survey on deep learning in medical image analysis. Medical Image Analysis 42, p. 60–88. Cited by: §1. [14] R. Miotto, F. Wang, S. Wang, X. Jiang, and J. Dudley (2018) Deep learning for healthcare: review, opportunities and challenges. Briefings in Bioinformatics 19 (6), p. 1236–1246. Cited by: §1. [15] OpenAI (2023) GPT-4 technical report. arXiv preprint arXiv:2303.08774. Cited by: §1. [16] A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, et al. (2021) Learning transferable visual models from natural language supervision. In Proceedings of ICML, Cited by: §2, §4.1. [17] P. Rajpurkar, J. Irvin, K. Zhu, B. Yang, H. Mehta, et al. (2017) CheXNet: radiologist-level pneumonia detection on chest x-rays with deep learning. arXiv preprint arXiv:1711.05225. Cited by: §1. [18] R. Tang et al. (2023) MedRAG: retrieval-augmented generation for medicine. arXiv preprint arXiv:2312.10912. Cited by: §2. [19] E. Topol (2019) High-performance medicine: the convergence of human and artificial intelligence. Nature Medicine 25, p. 44–56. Cited by: §2, §8. [20] Y. Wang et al. (2023) A survey on multimodal learning in medical imaging: progress, challenges, and future directions. Nature Machine Intelligence. Cited by: §9. [21] Z. Zhou et al. (2023) Foundation models for generalist medical artificial intelligence. Nature. Cited by: §2, §8.