ranjanhr1's picture
Add NayanaBench rendered dataset with 22 language splits
0892d53 verified
---
language:
- en
- hi
- kn
- ta
- te
- mr
- pa
- bn
- or
- ml
- gu
- sa
- ja
- ko
- zh
- de
- fr
- it
- ru
- ar
- es
- th
multilinguality: multilingual
task_categories:
- image-to-text
- object-detection
pretty_name: NayanaBench Rendered Dataset
size_categories:
- 1K<n<10K
tags:
- ocr
- text-rendering
- multilingual
- document-analysis
---
# NayanaBench Rendered Dataset
## Dataset Description
This dataset contains 2D rendered text images for document analysis and OCR tasks across 22 languages.
Each language split contains document images with text rendered in bounding boxes using authentic fonts.
### Dataset Statistics
- **Total Samples**: 4,400
- **Languages**: 22
- **Splits**: 22 language-based splits
### Language Splits
| Language | Code | Samples |
|----------|------|---------|
| EN | `en` | 200 |
| HI | `hi` | 200 |
| KN | `kn` | 200 |
| TA | `ta` | 200 |
| TE | `te` | 200 |
| MR | `mr` | 200 |
| PA | `pa` | 200 |
| BN | `bn` | 200 |
| OR | `or` | 200 |
| ML | `ml` | 200 |
| GU | `gu` | 200 |
| SA | `sa` | 200 |
| JA | `ja` | 200 |
| KO | `ko` | 200 |
| ZH | `zh` | 200 |
| DE | `de` | 200 |
| FR | `fr` | 200 |
| IT | `it` | 200 |
| RU | `ru` | 200 |
| AR | `ar` | 200 |
| ES | `es` | 200 |
| TH | `th` | 200 |
## Dataset Structure
### Data Fields
- `image`: The rendered document image
- `image_id`: Unique identifier for the sample
- `language`: Language code (ISO 639-1)
- `font_used`: Font file used for rendering
- `num_regions`: Number of text regions in the image
- `regions`: JSON string containing region metadata (bounding boxes, text content)
### Data Splits
This dataset has 22 splits, one per language. Each split can be loaded separately:
```python
from datasets import load_dataset
# Load a specific language
dataset = load_dataset("ranjanhr1/nayana_rendered", split="en")
# Load multiple languages
dataset_dict = load_dataset("ranjanhr1/nayana_rendered")
# Load specific languages
en_data = load_dataset("ranjanhr1/nayana_rendered", split="en")
hi_data = load_dataset("ranjanhr1/nayana_rendered", split="hi")
```
## Dataset Creation
This dataset was created by rendering text onto document images using authentic fonts for each language.
The rendering process uses:
- Tight canvas rendering (exact bounding box size)
- Optimal font size calculation
- Multi-line text wrapping
- Centered text alignment
### Source Dataset
Based on [Nayana-cognitivelab/NayanaBench](https://huggingface.co/datasets/Nayana-cognitivelab/NayanaBench)
## Citation
If you use this dataset, please cite:
```bibtex
@dataset{nayana_rendered,
title={NayanaBench Rendered Dataset},
author={Your Name},
year={2025},
publisher={Hugging Face},
url={https://huggingface.co/datasets/ranjanhr1/nayana_rendered}
}
```
## License
Please refer to the original NayanaBench dataset for licensing information.