| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - THUdyh/Oryx-Image-Data |
| | base_model: |
| | - Qwen/Qwen2-7B-Instruct |
| | pipeline_tag: text-generation |
| | language: |
| | - en |
| | - zh |
| | --- |
| | # Oryx-7B-Image |
| |
|
| | ## Model Summary |
| |
|
| | The Oryx-Image models are 7/34B parameter models trained on [Oryx-Image-Data](https://huggingface.co/datasets/THUdyh/Oryx-Image-Data), based on Qwen2 language model with a context window of 32K tokens. |
| |
|
| | Oryx offers an on-demand solution to seamlessly and efficiently process visual inputs with arbitrary spatial sizes and temporal lengths. |
| |
|
| | - **Repository:** https://github.com/Oryx-mllm/Oryx |
| | - **Languages:** English, Chinese |
| | - **Paper:** https://arxiv.org/abs/2409.12961 |
| |
|
| |
|
| | ### Model Architecture |
| |
|
| | - **Architecture:** Pre-trained [Oryx-ViT](https://huggingface.co/THUdyh/Oryx-ViT) + Qwen2-7B |
| | - **Data:** a mixture of 4M image data |
| | - **Precision:** BFloat16 |
| |
|
| | #### Hardware & Software |
| |
|
| | - **Hardware:** 64 * NVIDIA Tesla A100 |
| | - **Orchestration:** HuggingFace Trainer |
| | - **Code:** Pytorch |
| |
|
| | ## Citation |