Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,7 @@ tags:
|
|
| 12 |
|
| 13 |
# DatologyAI CLIP Classification Optimized ViT-B/32
|
| 14 |
|
| 15 |
-
**DatologyAI CLIP** is a state-of-the-art vision-language model that achieves superior performance through advanced data curation alone, without any architectural modifications. This classification-optimized ViT-B/32 model outperforms SigLIP2, MetaCLIP, and DFN on zero-shot classification benchmarks.
|
| 16 |
|
| 17 |
## Model Description
|
| 18 |
|
|
@@ -96,7 +96,7 @@ The model uses standard CLIP training objectives with no architectural modificat
|
|
| 96 |
|
| 97 |
## Training Data
|
| 98 |
|
| 99 |
-
The model was trained on 13B image-text
|
| 100 |
|
| 101 |
## Evaluation Results
|
| 102 |
|
|
@@ -136,7 +136,7 @@ The model was trained on 13B image-text pairs curated from the **DataComp Extra-
|
|
| 136 |
- **Weight decay:** 0.1
|
| 137 |
- **Batch size:** 32,768
|
| 138 |
- **Training samples:** 13B image-text pairs
|
| 139 |
-
- **Hardware:** Distributed training on
|
| 140 |
|
| 141 |
## Citation
|
| 142 |
|
|
|
|
| 12 |
|
| 13 |
# DatologyAI CLIP Classification Optimized ViT-B/32
|
| 14 |
|
| 15 |
+
**DatologyAI CLIP** is a state-of-the-art contrastive vision-language model that achieves superior performance through advanced data curation alone, without any architectural or training modifications. This classification-optimized ViT-B/32 model outperforms SigLIP2, MetaCLIP, and DFN on zero-shot classification benchmarks.
|
| 16 |
|
| 17 |
## Model Description
|
| 18 |
|
|
|
|
| 96 |
|
| 97 |
## Training Data
|
| 98 |
|
| 99 |
+
The model was trained on 13B image-text (multi-epoch) curated from the **DataComp-XL** dataset using DatologyAI's proprietary curation pipeline. The curation process selected high-quality, classification-relevant subsets from the 10B available pairs in DataComp-XL.
|
| 100 |
|
| 101 |
## Evaluation Results
|
| 102 |
|
|
|
|
| 136 |
- **Weight decay:** 0.1
|
| 137 |
- **Batch size:** 32,768
|
| 138 |
- **Training samples:** 13B image-text pairs
|
| 139 |
+
- **Hardware:** Distributed training on H100 GPUs
|
| 140 |
|
| 141 |
## Citation
|
| 142 |
|