Model Description
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct, trained for πevidence relevance classification or evidence filteringπ in medical RAG pipelines.
Given a clinical query and a candidate passage, the model outputs βYesβ if the passage contains supporting evidence and βNoβ otherwise.
This lightweight classifier is designed to help researchers:
- Improve retrieval quality in medical RAG systems.
- Filter irrelevant passages before generation.
- Build more reliable, interpretable RAG pipelines for medical QA.
For additional context, methodology, and full experimental details, please refer to our paper below.
Quick Start
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "Yale-BIDS-Chen/Llama-3.1-8B-Evidence-Filtering"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
# Instruction used during training
INSTRUCTION = (
"Given a query and a text passage, determine whether the passage contains supporting evidence for the query. "
"Supporting evidence means that the passage provides clear, relevant, and factual information that directly backs or justifies the answer to the query.\n\n"
"Respond with one of the following labels:\n\"Yes\" if the passage contains supporting evidence for the query.\n"
"\"No\" if the passage does not contain supporting evidence.\n"
"You should respond with only the label (Yes or No) without any additional explanation."
)
# Example query + retrieved passage
query = "What is the first-line treatment for acute angle-closure glaucoma?"
doc = "Acute angle-closure glaucoma requires immediate treatment with topical beta-blockers, alpha agonists, and systemic carbonic anhydrase inhibitors."
# Build chat-style prompt
content = tokenizer.apply_chat_template(
[
{"role": "system", "content": INSTRUCTION},
{"role": "user", "content": f"Question: {query}\nPassage: {doc}"}
],
add_generation_prompt=True,
tokenize=False,
)
# Tokenize
input_ids = tokenizer(content, return_tensors="pt").input_ids.to(model.device)
# Define stopping tokens (Llama-3 style)
terminators = [
tokenizer.eos_token_id,
tokenizer.convert_tokens_to_ids("<|eot_id|>")
]
# Generate evidence-filtering judgment
outputs = model.generate(
input_ids=input_ids,
max_new_tokens=256,
eos_token_id=terminators,
do_sample=False,
temperature=0.0,
)
# Decode model response
response = outputs[0][input_ids.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))
Training Setup
- Dataset: 3,200 queryβpassage pairs with expert-provided Yes/No labels (dataset to be released in a future update).
- Task: Given a query and a candidate passage, the model generates "Yes" if the passage contains supporting evidence and "No" otherwise.
- Objective: Causal language modeling (cross-entropy next-token loss).
- Prompt: See the Quick Start section for an example usage prompt.
- Hyperparameter Tuning: Five-fold cross-validation.
- Final Hyperparameters:
- Learning rate: 2e-6
- Batch size: 8
- Epochs: 3
- Training Framework: LLaMA-Factory.
Performance
Evaluation was conducted on 3,200 expert-annotated queryβpassage pairs using five-fold cross-validation.
| Model | Precision | Recall | F1 |
|---|---|---|---|
| Llama-3.1-8B (zero-shot) | 0.483 | 0.566 | 0.521 |
| GPT-4o (zero-shot) | 0.697 | 0.324 | 0.442 |
| Llama-3.1-8B (fine-tuned, ours) | 0.592 | 0.657 | 0.623 |
π₯ Fine-tuning yields substantial gains over all zero-shot baselines.
Intended Use
This model is intended for research purposes only.
Reference
Please see the information below to cite our paper.
@article{kim2025rethinking,
title={Rethinking Retrieval-Augmented Generation for Medicine: A Large-Scale, Systematic Expert Evaluation and Practical Insights},
author={Kim, Hyunjae and Sohn, Jiwoong and Gilson, Aidan and Cochran-Caggiano, Nicholas and Applebaum, Serina and Jin, Heeju and Park, Seihee and Park, Yujin and Park, Jiyeong and Choi, Seoyoung and others},
journal={arXiv preprint arXiv:2511.06738},
year={2025}
}
Contact
Feel free to email [email protected] if you have any questions.
- Downloads last month
- 38
Model tree for Yale-BIDS-Chen/Llama-3.1-8B-Evidence-Filtering
Base model
meta-llama/Llama-3.1-8B