DeepSeek-OCR GGUF (for llama.cpp PR #17400)

This repository provides GGUF model files required for running the DeepSeek-OCR / MTMD support introduced in the following llama.cpp pull request:

👉 llama.cpp PR #17400 https://github.com/ggml-org/llama.cpp/pull/17400

These models are only compatible with the PR branch and will not run on upstream llama.cpp main.


📥 Download

You can download the model files directly using:

huggingface-cli download <this-repo> --include "deepseek-ocr-f16.gguf" --local-dir gguf_models/deepseek-ai
huggingface-cli download <this-repo> --include "mmproj-deepseek-ocr-f16.gguf" --local-dir gguf_models/deepseek-ai

🛠 Build llama.cpp PR Branch

Clone llama.cpp and check out the PR branch:

git clone https://github.com/ggml-org/llama.cpp
cd llama.cpp

# Checkout PR branch (recommended: GitHub CLI)
gh pr checkout 17400
# or manually:
# git fetch origin pull/17400/head:pr17400
# git checkout pr17400

Build:

cmake -B build -DCMAKE_BUILD_TYPE=Release
cmake --build build -j

🚀 Run Example

Use the llama-mtmd-cli executable from the PR:

build/bin/llama-mtmd-cli \
  -m gguf_models/deepseek-ai/deepseek-ocr-f16.gguf \
  --mmproj gguf_models/deepseek-ai/mmproj-deepseek-ocr-f16.gguf \
  --image tmp/mtmd_test_data/Deepseek-OCR-2510.18234v1_page1.png \
  -p "<|grounding|>Convert the document to markdown." \
  --chat-template deepseek
Downloads last month
482
GGUF
Model size
3B params
Architecture
deepseek2-ocr
Hardware compatibility
Log In to view the estimation

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sabafallah/DeepSeek-OCR-GGUF

Quantized
(8)
this model