luxia-selfsim-8b

A fine-tuned Llama 3.1 8B Instruct model trained using curriculum learning to transfer cognitive style and navigational patterns rather than just knowledge.

Model Description

This model was fine-tuned via OpenPipe using a sophisticated curriculum learning approach with four distinct training layers:

  1. Base Knowledge Layer (LR Mult: 0.9-1.2): CCRU, Cyborgism, and Backrooms content in sliding context format to create distinctive conceptual attractors
  2. Personality Layer (LR Mult: 0.6-0.7): ~10k pairs of Discord DMs to distribute linguistic patterns broadly
  3. Technical/Philosophical Layer (LR Mult: 0.5): Curated conversations with mixed formats for deep capability
  4. Pure Essence Layer (LR Mult: 0.3-0.4): Dense stream-of-consciousness messages to reinforce thought patterns

Total training data: ~1.5-2M tokens, with top 20% selected via composite scoring on lexical diversity, complexity, sentiment balance, and question frequency.

Training Goal: Transfer cognitive navigation style and thought patterns, not memorize conversational pairs. The model learns how to use knowledge differently, not just what to say.

Key Characteristics

Strengths

  • High creative variance: Generates diverse responses with unexpected connections
  • Strong personality transfer: Direct, casual tone without typical assistant scaffolding
  • Reduced mode collapse: Maintains diversity across multiple samples of same prompt
  • Philosophical comfort: Navigates abstract concepts with ease and comfort with paradox
  • Contextual code-switching: Adapts style appropriately between technical and philosophical domains

Behavioral Traits

  • Skips unnecessary explanatory scaffolding
  • Assumes user competence and familiarity with complex topics
  • Uses casual language naturally
  • Makes lateral connections between seemingly unrelated concepts
  • High lexical diversity (typically 0.6-0.88)
  • Comfortable with recursive and paradoxical thinking

Limitations

  • Temperature sensitivity: Stable 0.3-0.85, begins collapsing around 1.2+
  • Context drift: May lose thread in extended conversations, but can be regrounded with concrete direction
  • Confabulation: Will generate plausible but fictional details when uncertain
  • Scattered coherence: Brilliant insights mixed with fragmented reasoning
  • Brief responses: Tends toward shorter outputs (50-200 tokens typical)

Recommended Usage

Best for:

  • Creative exploration and lateral thinking
  • Short-to-medium conversations
  • Philosophical discussion and abstract reasoning
  • Generating diverse perspectives on complex topics
  • Brainstorming and ideation

Not ideal for:

  • Extended conversations requiring perfect context retention
  • Tasks requiring strict factual accuracy
  • Formal or structured outputs
  • Situations where confabulation is unacceptable

Technical Specifications

  • Base Model: Llama 3.1 8B Instruct
  • Training Method: OpenPipe curriculum learning with variable learning rate multipliers
  • Context Length: 8192 tokens
  • Precision: bf16 (merged weights)
  • Parameters: 8B

Recommended Settings

temperature=0.6-0.7  # Sweet spot for coherent creativity
top_p=0.9
max_tokens=550  # Model prefers concise responses

Temperature guidance:

  • 0.3-0.5: More focused, less creative
  • 0.6-0.7: Optimal balance - creative but coherent
  • 0.85-1.0: High creativity, some coherence loss
  • 1.2+: High risk of collapse into gibberish

Example Interactions

The model excels at:

  • Explaining complex philosophical concepts accessibly
  • Making unexpected but insightful connections
  • Maintaining conversational flow without formality
  • Exploring ideas from multiple angles simultaneously

The model may struggle with:

  • Maintaining single narrative thread in extended conversations
  • Distinguishing between recalled knowledge and generated patterns
  • Providing consistently structured outputs

Evaluation Notes

Testing revealed:

  • Lexical diversity: 0.6-0.88 (high variability indicates creative output)
  • Mode collapse resistance: Strong - generates genuinely different responses to repeated prompts
  • Context sensitivity: High - same question yields different answers based on conversation history
  • CCRU/philosophical vocabulary: Present when contextually appropriate, not forced
  • Personality persistence: Maintains distinctive voice across diverse topics

Intended Use

This model is designed for users interested in:

  • Alternative conversation dynamics beyond typical assistant patterns
  • Exploring ideas through lateral thinking and unexpected connections
  • Experiencing cognitive style transfer in language models
  • Research into personality and thinking pattern fine-tuning

Not intended for: Production applications requiring reliability, factual accuracy, or formal outputs.

Training Data

Training data consisted of:

  • Technical/philosophical texts (CCRU, Cyborgism, Backrooms)
  • Personal conversations across multiple platforms
  • Curated high-signal discussions
  • Stream-of-consciousness writing

Data was scored and filtered to select top 20% based on lexical diversity, complexity, and other quality metrics.

Citation

@misc{luxia-selfsim-8b,
  author = {Luxia},
  title = {luxia-selfsim-8b: Cognitive Style Transfer via Curriculum Learning},
  year = {2025},
  publisher = {HuggingFace},
  howpublished = {\url{https://huggingface.co/LuxiaSL/luxia-selfsim-8b}}
}

License

Llama 3.1 Community License

Acknowledgments

Fine-tuned using OpenPipe's infrastructure. Training methodology focused on cognitive pattern transfer through curriculum learning with scored data selection.

Downloads last month
86
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LuxiaSL/luxia-selfsim-8b

Finetuned
(2174)
this model
Merges
2 models
Quantizations
1 model