Nano1337 commited on
Commit
e9914e9
·
verified ·
1 Parent(s): cc87b07

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
 
13
  # DatologyAI CLIP Classification Optimized ViT-B/32
14
 
15
- **DatologyAI CLIP** is a state-of-the-art vision-language model that achieves superior performance through advanced data curation alone, without any architectural modifications. This classification-optimized ViT-B/32 model outperforms SigLIP2, MetaCLIP, and DFN on zero-shot classification benchmarks.
16
 
17
  ## Model Description
18
 
@@ -96,7 +96,7 @@ The model uses standard CLIP training objectives with no architectural modificat
96
 
97
  ## Training Data
98
 
99
- The model was trained on 13B image-text pairs curated from the **DataComp Extra-Large** dataset using DatologyAI's proprietary curation pipeline. The curation process selected high-quality, classification-relevant subsets from the 10B available pairs in DataComp-XL.
100
 
101
  ## Evaluation Results
102
 
@@ -136,7 +136,7 @@ The model was trained on 13B image-text pairs curated from the **DataComp Extra-
136
  - **Weight decay:** 0.1
137
  - **Batch size:** 32,768
138
  - **Training samples:** 13B image-text pairs
139
- - **Hardware:** Distributed training on A100 GPUs
140
 
141
  ## Citation
142
 
 
12
 
13
  # DatologyAI CLIP Classification Optimized ViT-B/32
14
 
15
+ **DatologyAI CLIP** is a state-of-the-art contrastive vision-language model that achieves superior performance through advanced data curation alone, without any architectural or training modifications. This classification-optimized ViT-B/32 model outperforms SigLIP2, MetaCLIP, and DFN on zero-shot classification benchmarks.
16
 
17
  ## Model Description
18
 
 
96
 
97
  ## Training Data
98
 
99
+ The model was trained on 13B image-text (multi-epoch) curated from the **DataComp-XL** dataset using DatologyAI's proprietary curation pipeline. The curation process selected high-quality, classification-relevant subsets from the 10B available pairs in DataComp-XL.
100
 
101
  ## Evaluation Results
102
 
 
136
  - **Weight decay:** 0.1
137
  - **Batch size:** 32,768
138
  - **Training samples:** 13B image-text pairs
139
+ - **Hardware:** Distributed training on H100 GPUs
140
 
141
  ## Citation
142