π§ dnai-humour-0.5B-instruct
A lightweight, fast, and surprisingly witty instruction-tuned language model fine-tuned on curated OpenAssistant conversations. Built to respond clearly, efficiently, and with a touch of humor β without pretending to be a superintelligence.
π Overview
dnai-humour-0.5B-instruct is a fine-tuned variant of Qwen2.5-0.5B-Instruct, trained using a carefully selected subset of the OpenAssistant v1 dataset.
The focus is instruction following, conversational clarity, low-latency responses, and efficient deployment on modest hardware.
This model is small, fast, and does its job without unnecessary drama.
π― Main Capabilities
- π§Ύ Instruction following
- π¬ Conversational AI & chatbots
- π§ Reasonable reasoning (for 0.5B β letβs stay honest)
- π Light humor & friendly tone
- β‘ Fast inference and low memory usage
- π₯οΈ Suitable for edge devices & low-resource systems
π§ Model Details
| Item | Description |
|---|---|
| Base Model | Qwen2.5-0.5B-Instruct |
| Model Type | Decoder-only Transformer |
| Parameters | ~0.5 Billion |
| Fine-Tuning Method | Supervised Fine-Tuning (SFT) |
| Frameworks | PyTorch, Hugging Face Transformers, TRL |
| Precision Support | FP16 / INT8 (quantization-friendly) |
π Dataset
OpenAssistant v1 (OASST1)
- Source: OpenAssistant Project
- Type: Human-written multi-turn conversations
- Domains:
- Question answering
- Reasoning
- Coding help
- General knowledge
- Casual chat
π’ Data Used for Fine-Tuning
- Subset Size: ~15,000 conversations (smallest curated split)
- Selection Goal:
- High-quality instruction-response pairs
- Reduced noise
- Faster convergence
- Better alignment per token
Less data, more discipline.
β‘ Performance & Efficiency
- π Fast inference due to small parameter size
- π§ Low VRAM usage (runs comfortably on consumer GPUs)
- π¦ Easy to deploy on:
- Google Colab
- Lightning AI
- Local machines
- Edge setups
This model wonβt melt your GPU or your patience.
π Personality & Humor
- Polite, friendly, and occasionally funny
- Avoids being robotic when possible
- Does not hallucinate confidence like it knows everything
- Knows when to explain and when to shut up
Basically: helpful, not annoying.
π« Limitations
- Not designed for:
- Medical or legal advice
- High-stakes reasoning
- Large-context document analysis
- Still a 0.5B model β expectations should match reality
Small brain, well-trained.
π οΈ Intended Use Cases
- Educational chatbots
- Personal AI assistants
- Instruction-based tools
- Lightweight LLM experiments
- Fine-tuning & research demos
π License & Ethics
- Base model and dataset licenses apply
- Trained on publicly available, human-generated data
- No intentional harmful or restricted content
Use responsibly. Donβt blame the model for human mistakes.
π§ͺ Training Note
This model was fine-tuned using a minimal but high-quality dataset to balance performance and efficiency.
The goal was alignment per token, not brute-force scaling.
Quality > Quantity.
π€ Author
Fine-tuned by DarkNeuronAI
Built by a student. Powered by curiosity.
Optimized because resources are expensive.
β Final Words
If you need a small, fast, instruction-following model that doesnβt pretend to be GPT-4 β this one knows its place and performs it well.
- Downloads last month
- 4