File size: 1,024 Bytes
abca7bd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 |
# light_gpt_text_generator
## Overview
`light_gpt_text_generator` is a distilled, 6-layer generative Transformer based on the GPT-2 architecture. It is optimized for low-latency text completion and creative writing tasks where computational resources are limited (Edge devices/Mobile).
## Model Architecture
- **Type**: Causal Language Model (Decoder-only)
- **Layers**: 6 Transformer blocks (Reduced from GPT-2 Base's 12).
- **Embedding Dim**: 768
- **Heads**: 12 Multi-head self-attention.
- **Tokenizer**: Byte-level Byte Pair Encoding (BPE).
## Intended Use
- Real-time autocomplete for code or prose.
- Creative writing assistance and brainstorming.
- Chatbot prototyping for specific domains.
## Limitations
- **Hallucination**: High tendency to generate factually incorrect information.
- **Coherence**: Difficulty maintaining logical consistency over very long passages (>500 words).
- **Safety**: Lack of an integrated RLHF layer means the model may generate toxic or biased content if prompted inappropriately. |