| | --- |
| | base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-32B |
| | license: mit |
| | pipeline_tag: text-generation |
| | library_name: furiosa-llm |
| | tags: |
| | - furiosa-ai |
| | --- |
| | # Model Overview |
| | - **Model Architecture:** Qwen2 |
| | - **Input:** Text |
| | - **Output:** Text |
| | - **Model Optimizations:** |
| | - **Context Length:** 32k tokens |
| | - Maximum Prompt Length: 32768 tokens |
| | - Maximum Generation Length: 32768 tokens |
| | - **Intended Use Cases:** Intended for commercial and non-commercial use. Same as [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B), this models is intended for assistant-like chat. |
| | - **Release Date:** 08/26/2025 |
| | - **Version:** v2025.3 |
| | - **License(s):** [MIT License](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B/blob/main/LICENSE) |
| | - **Supported Inference Engine(s):** Furiosa LLM |
| | - **Supported Hardware Compatibility:** FuriosaAI RNGD |
| | - **Preferred Operating System(s):** Linux |
| |
|
| |
|
| | ## Description: |
| | This model is the pre-compiled version of the [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B), |
| | which is an auto-regressive language model that uses an optimized transformer architecture. |
| |
|
| | ## Usage |
| |
|
| | To run this model with [Furiosa-LLM](https://developer.furiosa.ai/latest/en/furiosa_llm/intro.html), |
| | follow the example command below after |
| | [installing Furiosa-LLM and its prerequisites](https://developer.furiosa.ai/latest/en/getting_started/furiosa_llm.html#installing-furiosa-llm). |
| |
|
| | ```sh |
| | furiosa-llm serve furiosa-ai/DeepSeek-R1-Distill-Qwen-32B \ |
| | --reasoning-parser deepseek_r1 |
| | ``` |
| |
|