| Model Type | | text generation, instruction tuned, conversation, assistant |
|
| Use Cases |
| Areas: | | research, controlled scientific experiments |
|
| Applications: | | alignment, text generation |
|
| Primary Use Cases: | | Developing language models for low-resource languages |
|
| Limitations: | | Not intended for deployment, Not suitable for human-facing interactions, Limited to Brazilian Portuguese, Not suitable for translation |
|
|
| Additional Notes | | Not suitable for real-world deployment without further evaluations. |
|
| Supported Languages | | Portuguese (Brazilian Portuguese) |
|
| Training Details |
| Data Sources: | | Instruct-Aira Dataset version 2.0 |
|
| Data Volume: | |
| Methodology: | |
| Hardware Used: | |
|
| Responsible Ai Considerations |
| Fairness: | | This model can exhibit biases and generate toxic content due to the social and historical stereotypes present in the training data. |
|
| Mitigation Strategies: | | Users are advised to perform risk analysis and have human moderation when the model's interactions reach an audience. |
|
|
| Input Output |
| Input Format: | | Text input with special token markers. |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use adequate repetition penalty and temperature settings to minimize repetition and verbosity. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | TeenyTinyLlama-460m-Chat fine-tuned on the Instruct-Aira dataset and evaluated on multiple benchmarks. |
|
|
|