model_name_or_path: Qwen/Qwen2.5-3B-Instruct stage: sft finetuning_type: lora dataset: - unlearn dataset_dir: ./data output_dir: output/qwen-unlearn num_train_epochs: 6 max_steps: 30 per_device_train_batch_size: 1 gradient_accumulation_steps: 4 learning_rate: 3e-5 lora_rank: 64 lora_alpha: 64 lora_dropout: 0.05 lr_scheduler_type: cosine warmup_ratio: 0.05 fp16: true save_steps: 10 logging_steps: 1 report_to: none log_level: debug log_level_replica: debug disable_tqdm: false overwrite_output_dir: true packing: false do_train: true