Model Description

This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct, trained for πŸš€evidence relevance classification or evidence filteringπŸš€ in medical RAG pipelines.
Given a clinical query and a candidate passage, the model outputs β€œYes” if the passage contains supporting evidence and β€œNo” otherwise.

This lightweight classifier is designed to help researchers:

  • Improve retrieval quality in medical RAG systems.
  • Filter irrelevant passages before generation.
  • Build more reliable, interpretable RAG pipelines for medical QA.

For additional context, methodology, and full experimental details, please refer to our paper below.

πŸ“„ Paper: Rethinking Retrieval-Augmented Generation for Medicine: A Large-Scale, Systematic Expert Evaluation and Practical Insights

Quick Start

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "Yale-BIDS-Chen/Llama-3.1-8B-Evidence-Filtering"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

# Instruction used during training
INSTRUCTION = (
    "Given a query and a text passage, determine whether the passage contains supporting evidence for the query. "
    "Supporting evidence means that the passage provides clear, relevant, and factual information that directly backs or justifies the answer to the query.\n\n"
    "Respond with one of the following labels:\n\"Yes\" if the passage contains supporting evidence for the query.\n"
    "\"No\" if the passage does not contain supporting evidence.\n"
    "You should respond with only the label (Yes or No) without any additional explanation."
)

# Example query + retrieved passage
query = "What is the first-line treatment for acute angle-closure glaucoma?"
doc = "Acute angle-closure glaucoma requires immediate treatment with topical beta-blockers, alpha agonists, and systemic carbonic anhydrase inhibitors."

# Build chat-style prompt
content = tokenizer.apply_chat_template(
    [
        {"role": "system", "content": INSTRUCTION},
        {"role": "user", "content": f"Question: {query}\nPassage: {doc}"}
    ],
    add_generation_prompt=True,
    tokenize=False,
)

# Tokenize
input_ids = tokenizer(content, return_tensors="pt").input_ids.to(model.device)

# Define stopping tokens (Llama-3 style)
terminators = [
    tokenizer.eos_token_id,
    tokenizer.convert_tokens_to_ids("<|eot_id|>")
]

# Generate evidence-filtering judgment
outputs = model.generate(
    input_ids=input_ids,
    max_new_tokens=256,
    eos_token_id=terminators,
    do_sample=False,
    temperature=0.0,
)

# Decode model response
response = outputs[0][input_ids.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))

Training Setup

  • Dataset: 3,200 query–passage pairs with expert-provided Yes/No labels (dataset to be released in a future update).
  • Task: Given a query and a candidate passage, the model generates "Yes" if the passage contains supporting evidence and "No" otherwise.
  • Objective: Causal language modeling (cross-entropy next-token loss).
  • Prompt: See the Quick Start section for an example usage prompt.
  • Hyperparameter Tuning: Five-fold cross-validation.
  • Final Hyperparameters:
    • Learning rate: 2e-6
    • Batch size: 8
    • Epochs: 3
  • Training Framework: LLaMA-Factory.

Performance

Evaluation was conducted on 3,200 expert-annotated query–passage pairs using five-fold cross-validation.

Model Precision Recall F1
Llama-3.1-8B (zero-shot) 0.483 0.566 0.521
GPT-4o (zero-shot) 0.697 0.324 0.442
Llama-3.1-8B (fine-tuned, ours) 0.592 0.657 0.623

πŸ”₯ Fine-tuning yields substantial gains over all zero-shot baselines.

Intended Use

This model is intended for research purposes only.

Reference

Please see the information below to cite our paper.

@article{kim2025rethinking,
  title={Rethinking Retrieval-Augmented Generation for Medicine: A Large-Scale, Systematic Expert Evaluation and Practical Insights},
  author={Kim, Hyunjae and Sohn, Jiwoong and Gilson, Aidan and Cochran-Caggiano, Nicholas and Applebaum, Serina and Jin, Heeju and Park, Seihee and Park, Yujin and Park, Jiyeong and Choi, Seoyoung and others},
  journal={arXiv preprint arXiv:2511.06738},
  year={2025}
}

Contact

Feel free to email [email protected] if you have any questions.

Downloads last month
38
Safetensors
Model size
8B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Yale-BIDS-Chen/Llama-3.1-8B-Evidence-Filtering

Finetuned
(2071)
this model