DhruvaBansal00
commited on
Commit
·
6306fb5
1
Parent(s):
18fdce6
Model naming
Browse files
README.md
CHANGED
|
@@ -11,9 +11,9 @@ We’re thrilled to introduce RefuelLLM-2 and RefuelLLM-2-small, the next versio
|
|
| 11 |
|
| 12 |
1. RefuelLLM-2 (83.82%) outperforms all state-of-the-art LLMs, including GPT-4-Turbo (80.88%), Claude-3-Opus (79.19%) and Gemini-1.5-Pro (74.59%), across a benchmark of ~30 data labeling tasks.
|
| 13 |
2. RefuelLLM-2 is a Mixtral-8x7B base model, trained on a corpus of 2750+ datasets, spanning tasks such as classification, reading comprehension, structured attribute extraction and entity resolution.
|
| 14 |
-
3. RefuelLLM-2-small (79.67%) outperforms all comparable LLMs including Claude3-Sonnet (70.99%), Haiku (69.23%) and GPT-3.5-Turbo (68.13%). The model was trained with the same recipe as RefuelLLM-2, but on top of Llama3-8B base.
|
| 15 |
|
| 16 |
-
As a part of this announcement, we are open-sourcing RefuelLLM-2-small
|
| 17 |
|
| 18 |
**Model developers** Refuel AI
|
| 19 |
|
|
|
|
| 11 |
|
| 12 |
1. RefuelLLM-2 (83.82%) outperforms all state-of-the-art LLMs, including GPT-4-Turbo (80.88%), Claude-3-Opus (79.19%) and Gemini-1.5-Pro (74.59%), across a benchmark of ~30 data labeling tasks.
|
| 13 |
2. RefuelLLM-2 is a Mixtral-8x7B base model, trained on a corpus of 2750+ datasets, spanning tasks such as classification, reading comprehension, structured attribute extraction and entity resolution.
|
| 14 |
+
3. RefuelLLM-2-small (79.67%), aka Llama-3-Refueled, outperforms all comparable LLMs including Claude3-Sonnet (70.99%), Haiku (69.23%) and GPT-3.5-Turbo (68.13%). The model was trained with the same recipe as RefuelLLM-2, but on top of Llama3-8B base.
|
| 15 |
|
| 16 |
+
As a part of this announcement, we are open-sourcing RefuelLLM-2-small for the community to build on top of.
|
| 17 |
|
| 18 |
**Model developers** Refuel AI
|
| 19 |
|