Update README.md
Browse files
README.md
CHANGED
|
@@ -24,57 +24,7 @@ datasets:
|
|
| 24 |
- OPUS
|
| 25 |
- HPLT
|
| 26 |
- Tatoeba
|
| 27 |
-
models:
|
| 28 |
-
- Helsinki-NLP
|
| 29 |
-
- QUICKMT
|
| 30 |
-
- Argos
|
| 31 |
-
- Lego-MT/Lego-MT
|
| 32 |
-
- HPLT
|
| 33 |
-
- HPLT-OPUS
|
| 34 |
-
- Google
|
| 35 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-mul-mul
|
| 36 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-mul-deu_eng_nld
|
| 37 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-mul-deu_eng_fra_por_spa
|
| 38 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-deu_eng_fra_por_spa-mul
|
| 39 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-roa-deu_eng_fra_por_spa
|
| 40 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-deu_eng_fra_por_spa-roa
|
| 41 |
-
- Helsinki-NLP/opus-mt-tc-bible-big-roa-en
|
| 42 |
-
- facebook/nllb-200-distilled-600M
|
| 43 |
-
- facebook/nllb-200-distilled-1.3B
|
| 44 |
-
- facebook/nllb-200-1.3B
|
| 45 |
-
- facebook/nllb-200-3.3B
|
| 46 |
-
- facebook/mbart-large-50-many-to-many-mmt
|
| 47 |
-
- facebook/mbart-large-50-one-to-many-mmt
|
| 48 |
-
- facebook/mbart-large-50-many-to-one-mmt
|
| 49 |
-
- facebook/m2m100_418M
|
| 50 |
-
- facebook/m2m100_1.2B
|
| 51 |
-
- bigscience/mt0-small
|
| 52 |
-
- bigscience/mt0-base
|
| 53 |
-
- bigscience/mt0-large
|
| 54 |
-
- bigscience/mt0-xl
|
| 55 |
-
- bigscience/bloomz-560m
|
| 56 |
-
- bigscience/bloomz-1b1
|
| 57 |
-
- bigscience/bloomz-1b7
|
| 58 |
-
- bigscience/bloomz-3b
|
| 59 |
-
- t5-small
|
| 60 |
-
- t5-base
|
| 61 |
-
- t5-large
|
| 62 |
-
- google/flan-t5-small
|
| 63 |
-
- google/flan-t5-base
|
| 64 |
-
- google/flan-t5-large
|
| 65 |
-
- google/flan-t5-xl
|
| 66 |
-
- google/madlad400-3b-mt
|
| 67 |
-
- jbochi/madlad400-3b-mt
|
| 68 |
-
- utter-project/EuroLLM-1.7B
|
| 69 |
-
- utter-project/EuroLLM-1.7B-Instruct
|
| 70 |
-
- Unbabel/Tower-Plus-2B
|
| 71 |
-
- Unbabel/TowerInstruct-7B-v0.2
|
| 72 |
-
- Unbabel/TowerInstruct-Mistral-7B-v0.2
|
| 73 |
-
- HuggingFaceTB/SmolLM3-3B
|
| 74 |
-
- winninghealth/WiNGPT-Babel-2
|
| 75 |
-
- tencent/Hunyuan-MT-7B
|
| 76 |
-
- openGPT-X/Teuken-7B-instruct-commercial-v0.4
|
| 77 |
-
- openGPT-X/Teuken-7B-instruct-v0.6
|
| 78 |
---
|
| 79 |
|
| 80 |
```text
|
|
|
|
| 24 |
- OPUS
|
| 25 |
- HPLT
|
| 26 |
- Tatoeba
|
| 27 |
+
models: ["Helsinki-NLP", "QUICKMT", "Argos", "Lego-MT/Lego-MT", "HPLT", "HPLT-OPUS", "Google", "Helsinki-NLP/opus-mt-tc-bible-big-mul-mul", "Helsinki-NLP/opus-mt-tc-bible-big-mul-deu_eng_nld", "Helsinki-NLP/opus-mt-tc-bible-big-mul-deu_eng_fra_por_spa", "Helsinki-NLP/opus-mt-tc-bible-big-deu_eng_fra_por_spa-mul", "Helsinki-NLP/opus-mt-tc-bible-big-roa-deu_eng_fra_por_spa", "Helsinki-NLP/opus-mt-tc-bible-big-deu_eng_fra_por_spa-roa", "Helsinki-NLP/opus-mt-tc-bible-big-roa-en", "facebook/nllb-200-distilled-600M", "facebook/nllb-200-distilled-1.3B", "facebook/nllb-200-1.3B", "facebook/nllb-200-3.3B", "facebook/mbart-large-50-many-to-many-mmt", "facebook/mbart-large-50-one-to-many-mmt", "facebook/mbart-large-50-many-to-one-mmt", "facebook/m2m100_418M", "facebook/m2m100_1.2B", "alirezamsh/small100", "bigscience/mt0-small", "bigscience/mt0-base", "bigscience/mt0-large", "bigscience/mt0-xl", "bigscience/bloomz-560m", "bigscience/bloomz-1b1", "bigscience/bloomz-1b7", "bigscience/bloomz-3b", "t5-small", "t5-base", "t5-large", "google/flan-t5-small", "google/flan-t5-base", "google/flan-t5-large", "google/flan-t5-xl", "google/madlad400-3b-mt", "Heng666/madlad400-3b-mt-ct2", "Heng666/madlad400-3b-mt-ct2-int8", "Heng666/madlad400-7b-mt-ct2-int8", "BSC-LT/salamandraTA-2b-instruct", "BSC-LT/salamandraTA-7b-instruct", "utter-project/EuroLLM-1.7B", "utter-project/EuroLLM-1.7B-Instruct", "Unbabel/Tower-Plus-2B", "Unbabel/TowerInstruct-7B-v0.2", "Unbabel/TowerInstruct-Mistral-7B-v0.2", "HuggingFaceTB/SmolLM3-3B", "winninghealth/WiNGPT-Babel-2-1", "winninghealth/WiNGPT-Babel-2", "winninghealth/WiNGPT-Babel", "tencent/Hunyuan-MT-7B", "openGPT-X/Teuken-7B-instruct-commercial-v0.4", "openGPT-X/Teuken-7B-instruct-v0.6"]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
---
|
| 29 |
|
| 30 |
```text
|