LimRank: Less is More for Reasoning-Intensive Information Reranking

This repository contains the limrank-7b model, based on Qwen2.5-7B, which was presented in the paper LimRank: Less is More for Reasoning-Intensive Information Reranking.

LimRank demonstrates an efficient approach to adapt modern Large Language Models (LLMs) for reasoning-intensive information reranking tasks. This is achieved by leveraging LIMRANK-SYNTHESIZER, a reusable and open-source pipeline that generates minimal yet high-quality synthetic supervision data. Through this approach, LimRank achieves competitive performance on challenging benchmarks like BRIGHT and FollowIR, utilizing less than 5% of the data typically required by prior methods. The model also shows strong generalization capabilities across various downstream tasks, including scientific literature search and retrieval-augmented generation.

Links

Citation

If you find our paper useful, please cite our work:

@misc{song2025limrankreasoningintensiveinformationreranking,
      title={LimRank: Less is More for Reasoning-Intensive Information Reranking}, 
      author={Tingyu Song and Yilun Zhao and Siyue Zhang and Chen Zhao and Arman Cohan},
      year={2025},
      eprint={2510.23544},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2510.23544}, 
}

Acknowledgements

We would like to thank the authors of the following papers and repos for their open-source contributions.

License

The model is released under the MIT License.

Downloads last month
10
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including songtingyu/limrank