LoRA: Low-Rank Adaptation of Large Language Models
Paper
β’
2106.09685
β’
Published
β’
57
google/flan-t5-base on a curated dataset of Stack Overflow programming questions. It was trained using LoRA (Low-Rank Adaptation) for parameter-efficient fine-tuning, making it compact, efficient, and effective at modeling developer-style Q&A tasks.
google/flan-t5-basepeftadapter_model.safetensorsadapter_config.jsonr: 8lora_alpha: 16lora_dropout: 0.1bias: "none"task_type: "SEQ_2_SEQ_LM"from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
from peft import PeftModel
# Load tokenizer and base model
tokenizer = AutoTokenizer.from_pretrained("google/flan-t5-base")
base_model = AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-base")
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "your-model-folder")
model.eval()
# Inference
prompt = "Rewrite this question more clearly: why is my javascript function undefined?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
π§ͺ Intended Use This model is best suited for: Code-aware chatbot assistants Prompt engineering for developer tools Developer-focused summarization / rephrasing Auto-moderation / clarification of tech questions
β οΈ Limitations Not trained for code generation or long-form answers May hallucinate incorrect or generic responses Finetuned only on Stack Overflow β domain-specific
β¨ Acknowledgements Hugging Face Transformers LoRA (PEFT) Stack Overflow for open data FLAN-T5: Scaling Instruction-Finetuned Models
π οΈ Created with love by Kunj | Model suggestion & guidance by ChatGPT
Base model
google/flan-t5-base