finetune-ec9dc043

Fine-tuned version of Qwen/Qwen3-1.7B using DPO.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("Canfield/finetune-ec9dc043")
tokenizer = AutoTokenizer.from_pretrained("Canfield/finetune-ec9dc043")

# Generate text
messages = [
    {"role": "user", "content": "Hello, how are you?"}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Training Details

  • Method: DPO
  • Base Model: Qwen/Qwen3-1.7B
  • LoRA merged and quantization removed for inference compatibility
Downloads last month
3
Safetensors
Model size
1B params
Tensor type
F32
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Canfield/finetune-ec9dc043

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(667)
this model