internlm-chatbode-20b

ChatBode Logo

O InternLm-ChatBode Γ© um modelo de linguagem ajustado para o idioma portuguΓͺs, desenvolvido a partir do modelo InternLM2. Este modelo foi refinado atravΓ©s do processo de fine-tuning utilizando o dataset UltraAlpaca.

CaracterΓ­sticas Principais

  • Modelo Base: internlm/internlm2-chat-20b
  • Dataset para Fine-tuning: UltraAlpaca
  • Treinamento: O treinamento foi realizado a partir do fine-tuning, usando QLoRA, do internlm2-chat-20b.

Exemplo de uso

A seguir um exemplo de cΓ³digo de como carregar e utilizar o modelo:

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("recogna-nlp/internlm-chatbode-20b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("recogna-nlp/internlm-chatbode-20b", torch_dtype=torch.float16, trust_remote_code=True).cuda()
model = model.eval()
response, history = model.chat(tokenizer, "OlΓ‘", history=[])
print(response)
response, history = model.chat(tokenizer, "O que Γ© o Teorema de PitΓ‘goras? Me dΓͺ um exemplo", history=history)
print(response)

As respostas podem ser geradas via stream utilizando o mΓ©todo stream_chat:

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_path = "recogna-nlp/internlm-chatbode-20b"
model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.float16, trust_remote_code=True).cuda()
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)

model = model.eval()
length = 0
for response, history in model.stream_chat(tokenizer, "OlΓ‘", history=[]):
    print(response[length:], flush=True, end="")
    length = len(response)

Open Portuguese LLM Leaderboard Evaluation Results

Detailed results can be found here and on the πŸš€ Open Portuguese LLM Leaderboard

Metric Value
Average 71.68
ENEM Challenge (No Images) 65.78
BLUEX (No Images) 58.69
OAB Exams 43.33
Assin2 RTE 91.53
Assin2 STS 78.95
FaQuAD NLI 81.36
HateBR Binary 81.72
PT Hate Speech Binary 73.66
tweetSentBR 70.11
Downloads last month
24
Safetensors
Model size
20B params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Space using recogna-nlp/internlm-chatbode-20b 1

Evaluation results