--- license: mit language: - pt pipeline_tag: text-generation base_model: - AxionLab-official/MiniBot-0.9M-Base library_name: transformers --- # ๐ง MiniBot-0.9M-Instruct > **Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.** [](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) [](https://opensource.org/licenses/MIT) [](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) [](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) --- ## ๐ Overview **MiniBot-0.9M-Instruct** is the instruction-tuned version of [MiniBot-0.9M-Base](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Base), designed to follow prompts more accurately, respond to user inputs, and generate more coherent conversational outputs in **Portuguese**. Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions. --- ## ๐ฏ Key Characteristics | Attribute | Detail | |---|---| | ๐ง๐ท **Language** | Portuguese (primary) | | ๐ง **Architecture** | GPT-2 style (Transformer decoder-only) | | ๐ค **Embeddings** | GPT-2 compatible | | ๐ **Parameters** | ~900K | | โ๏ธ **Base Model** | MiniBot-0.9M-Base | | ๐ฏ **Fine-tuning** | Instruction tuning (supervised) | | โ **Alignment** | Basic prompt-following behavior | --- ## ๐ง What Changed from Base? Instruction tuning introduced significant behavioral improvements with no architectural changes: | Feature | Base | Instruct | |---|---|---| | Prompt understanding | โ | โ | | Conversational flow | โ ๏ธ Partial | โ | | Instruction following | โ | โ | | Overall coherence | Low | Improved | | Practical usability | Experimental | Functional | > ๐ก The model is now significantly more usable in chat scenarios. --- ## ๐๏ธ Architecture The core architecture remains identical to the base model: - **Decoder-only Transformer** (GPT-2 style) - Token embeddings + positional embeddings - Self-attention + MLP blocks - Autoregressive generation No structural changes were made โ only behavioral improvement through fine-tuning. --- ## ๐ Fine-Tuning Dataset The model was fine-tuned on a Portuguese instruction-style conversational dataset composed of: - ๐ฌ Questions and answers - ๐ Simple instructions - ๐ค Assistant-style chat - ๐ญ Basic roleplay - ๐ฃ๏ธ Natural conversations **Expected format:** ``` User: Me explique o que รฉ gravidade Bot: A gravidade รฉ a forรงa que atrai objetos com massa... ``` **Training strategy:** - Supervised Fine-Tuning (SFT) - Pattern learning for instruction-following - No RLHF or preference optimization --- ## ๐ก Capabilities ### โ Strengths - Following simple instructions - Answering basic questions - Conversing more naturally - Higher coherence in short responses - More consistent dialogue structure ### โ Limitations - Reasoning is still limited - May generate incorrect facts - Does not retain long context - Sensitive to poorly structured prompts > โ ๏ธ Even with instruction tuning, this remains an extremely small model. Adjust expectations accordingly. --- ## ๐ Getting Started ### Installation ```bash pip install transformers torch ``` ### Usage with Hugging Face Transformers ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "AxionLab-official/MiniBot-0.9M-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) prompt = "User: Me diga uma curiosidade sobre o espaรงo\nBot:" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate( **inputs, max_new_tokens=80, temperature=0.7, top_p=0.9, do_sample=True, ) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ### โ๏ธ Recommended Settings | Parameter | Recommended Value | Description | |---|---|---| | `temperature` | `0.6 โ 0.8` | Controls randomness | | `top_p` | `0.85 โ 0.95` | Nucleus sampling | | `do_sample` | `True` | Enable sampling | | `max_new_tokens` | `40 โ 100` | Response length | > ๐ก Instruct models tend to perform better at lower temperatures. Try values around `0.65` for more accurate and focused responses. --- ## ๐งช Intended Use Cases | Use Case | Suitability | |---|---| | ๐ฌ Lightweight Portuguese chatbots | โ Ideal | | ๐ฎ NPCs and games | โ Ideal | | ๐ง Fine-tuning experiments | โ Ideal | | ๐ NLP education | โ Ideal | | โก Local / CPU-only applications | โ Ideal | | ๐ญ Critical production environments | โ Not recommended | --- ## โ ๏ธ Disclaimer - Extremely small model (~900K parameters) - No robust alignment (no RLHF) - May generate incorrect or nonsensical responses - **Not suitable for critical production environments** --- ## ๐ฎ Future Work - [ ] ๐ง Reasoning-tuned version (`MiniBot-Reason`) - [ ] ๐ Scaling to 1Mโ10M parameters - [ ] ๐ Larger and more diverse dataset - [ ] ๐ค Improved response alignment - [ ] ๐งฉ Tool-use experiments --- ## ๐ License Distributed under the **MIT License**. See [`LICENSE`](LICENSE) for more details. --- ## ๐ค Author Developed by **[AxionLab](https://huggingface.co/AxionLab-official)** ๐ฌ ---