Base GPT-2 XL Better SFT โ€” HF-Only Assistant Coherence Continuation

Continuation checkpoint trained from Darkstorm1826/base-gpt2xl-better-sft-experimental using a JAX/Optax TPU full fine-tune pass.

This experimental continuation uses /kaggle/working/distilled.jsonl, built only from Hugging Face datasets.

Focus:

  • broader assistant behavior
  • chat/instruction following
  • creative writing
  • everyday procedural answers
  • context-grounded QA
  • truthfulness and hallucination reduction
  • cleaner stopping behavior

Prompt format:

<|system|> You are a helpful, honest, and careful AI assistant.<|end|> <|user|> Question here<|end|> <|assistant|> Answer here<|end|>

Downloads last month
38
Safetensors
Model size
2B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Darkstorm1826/base-gpt2xl-better-sft-experimental

Unable to build the model tree, the base model loops to the model itself. Learn more.