--- license: mit language: - pt pipeline_tag: text-generation base_model: - AxionLab-official/MiniBot-0.9M-Base library_name: transformers --- # ๐Ÿง  MiniBot-0.9M-Instruct > **Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.** [![Model](https://img.shields.io/badge/๐Ÿค—%20Hugging%20Face-MiniBot--0.9M--Instruct-yellow)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) [![License](https://img.shields.io/badge/License-MIT-green.svg)](https://opensource.org/licenses/MIT) [![Language](https://img.shields.io/badge/Language-Portuguese-blue)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) [![Parameters](https://img.shields.io/badge/Parameters-~900K-orange)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct) --- ## ๐Ÿ“Œ Overview **MiniBot-0.9M-Instruct** is the instruction-tuned version of [MiniBot-0.9M-Base](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Base), designed to follow prompts more accurately, respond to user inputs, and generate more coherent conversational outputs in **Portuguese**. Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions. --- ## ๐ŸŽฏ Key Characteristics | Attribute | Detail | |---|---| | ๐Ÿ‡ง๐Ÿ‡ท **Language** | Portuguese (primary) | | ๐Ÿง  **Architecture** | GPT-2 style (Transformer decoder-only) | | ๐Ÿ”ค **Embeddings** | GPT-2 compatible | | ๐Ÿ“‰ **Parameters** | ~900K | | โš™๏ธ **Base Model** | MiniBot-0.9M-Base | | ๐ŸŽฏ **Fine-tuning** | Instruction tuning (supervised) | | โœ… **Alignment** | Basic prompt-following behavior | --- ## ๐Ÿง  What Changed from Base? Instruction tuning introduced significant behavioral improvements with no architectural changes: | Feature | Base | Instruct | |---|---|---| | Prompt understanding | โŒ | โœ… | | Conversational flow | โš ๏ธ Partial | โœ… | | Instruction following | โŒ | โœ… | | Overall coherence | Low | Improved | | Practical usability | Experimental | Functional | > ๐Ÿ’ก The model is now significantly more usable in chat scenarios. --- ## ๐Ÿ—๏ธ Architecture The core architecture remains identical to the base model: - **Decoder-only Transformer** (GPT-2 style) - Token embeddings + positional embeddings - Self-attention + MLP blocks - Autoregressive generation No structural changes were made โ€” only behavioral improvement through fine-tuning. --- ## ๐Ÿ“š Fine-Tuning Dataset The model was fine-tuned on a Portuguese instruction-style conversational dataset composed of: - ๐Ÿ’ฌ Questions and answers - ๐Ÿ“‹ Simple instructions - ๐Ÿค– Assistant-style chat - ๐ŸŽญ Basic roleplay - ๐Ÿ—ฃ๏ธ Natural conversations **Expected format:** ``` User: Me explique o que รฉ gravidade Bot: A gravidade รฉ a forรงa que atrai objetos com massa... ``` **Training strategy:** - Supervised Fine-Tuning (SFT) - Pattern learning for instruction-following - No RLHF or preference optimization --- ## ๐Ÿ’ก Capabilities ### โœ… Strengths - Following simple instructions - Answering basic questions - Conversing more naturally - Higher coherence in short responses - More consistent dialogue structure ### โŒ Limitations - Reasoning is still limited - May generate incorrect facts - Does not retain long context - Sensitive to poorly structured prompts > โš ๏ธ Even with instruction tuning, this remains an extremely small model. Adjust expectations accordingly. --- ## ๐Ÿš€ Getting Started ### Installation ```bash pip install transformers torch ``` ### Usage with Hugging Face Transformers ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "AxionLab-official/MiniBot-0.9M-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) prompt = "User: Me diga uma curiosidade sobre o espaรงo\nBot:" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate( **inputs, max_new_tokens=80, temperature=0.7, top_p=0.9, do_sample=True, ) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ### โš™๏ธ Recommended Settings | Parameter | Recommended Value | Description | |---|---|---| | `temperature` | `0.6 โ€“ 0.8` | Controls randomness | | `top_p` | `0.85 โ€“ 0.95` | Nucleus sampling | | `do_sample` | `True` | Enable sampling | | `max_new_tokens` | `40 โ€“ 100` | Response length | > ๐Ÿ’ก Instruct models tend to perform better at lower temperatures. Try values around `0.65` for more accurate and focused responses. --- ## ๐Ÿงช Intended Use Cases | Use Case | Suitability | |---|---| | ๐Ÿ’ฌ Lightweight Portuguese chatbots | โœ… Ideal | | ๐ŸŽฎ NPCs and games | โœ… Ideal | | ๐Ÿง  Fine-tuning experiments | โœ… Ideal | | ๐Ÿ“š NLP education | โœ… Ideal | | โšก Local / CPU-only applications | โœ… Ideal | | ๐Ÿญ Critical production environments | โŒ Not recommended | --- ## โš ๏ธ Disclaimer - Extremely small model (~900K parameters) - No robust alignment (no RLHF) - May generate incorrect or nonsensical responses - **Not suitable for critical production environments** --- ## ๐Ÿ”ฎ Future Work - [ ] ๐Ÿง  Reasoning-tuned version (`MiniBot-Reason`) - [ ] ๐Ÿ“ˆ Scaling to 1Mโ€“10M parameters - [ ] ๐Ÿ“š Larger and more diverse dataset - [ ] ๐Ÿค– Improved response alignment - [ ] ๐Ÿงฉ Tool-use experiments --- ## ๐Ÿ“œ License Distributed under the **MIT License**. See [`LICENSE`](LICENSE) for more details. --- ## ๐Ÿ‘ค Author Developed by **[AxionLab](https://huggingface.co/AxionLab-official)** ๐Ÿ”ฌ ---
MiniBot-0.9M-Instruct ยท AxionLab ยท MIT License