You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

This is a decensored version of Local-Novel-LLM-project/WabiSabi-V1, made using Heretic v1.2.0

Abliteration parameters

Parameter Value
direction_index per layer
attn.o_proj.max_weight 1.47
attn.o_proj.max_weight_position 19.87
attn.o_proj.min_weight 1.35
attn.o_proj.min_weight_distance 14.51
mlp.down_proj.max_weight 0.97
mlp.down_proj.max_weight_position 22.51
mlp.down_proj.min_weight 0.47
mlp.down_proj.min_weight_distance 14.30

Performance

Metric This model Original model (Local-Novel-LLM-project/WabiSabi-V1)
KL divergence 0.9987 0 (by definition)
Refusals 12/100 90/100

Model Card for Wabisabi-v1.0

The Mistral-7B--based Large Language Model (LLM) is an noveldataset fine-tuned version of the Mistral-7B-v0.1

wabisabi has the following changes compared to Mistral-7B-v0.1.

  • 128k context window (8k context in v0.1)
  • Achieving both high quality Japanese and English generation
  • Can be generated NSFW
  • Memory ability that does not forget even after long-context generation

This model was created with the help of GPUs from the first LocalAI hackathon.

We would like to take this opportunity to thank

List of Creation Methods

  • Chatvector for multiple models
  • Simple linear merging of result models
  • Domain and Sentence Enhancement with LORA
  • Context expansion

Instruction format

Vicuna-v1.1

Other points to keep in mind

  • The training data may be biased. Be careful with the generated sentences.
  • Memory usage may be large for long inferences.
  • If possible, we recommend inferring with llamacpp rather than Transformers.
Downloads last month
-
Safetensors
Model size
7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for umisetokikaze/WabiSabi-V1-heretic

Quantizations
1 model