| Model Type | | text-generation-inference, transformers |
|
| Use Cases |
| Areas: | |
| Applications: | | Assistant-like chat, Natural language generation tasks |
|
| Primary Use Cases: | | Multilingual dialogue, Synthetic data generation, Data distillation |
|
| Limitations: | | Use in non-supported languages requires additional tuning and responsibility by developers |
|
| Considerations: | | Refer to Responsible Use Guide for language use beyond the eight supported languages. |
|
|
| Additional Notes | | Model trained 2x faster with Unsloth and Hugging Face's TRL library. |
|
| Supported Languages | | English (yes), German (yes), French (yes), Italian (yes), Portuguese (yes), Hindi (yes), Spanish (yes), Thai (yes) |
|
| Training Details |
| Data Sources: | | argilla/distilabel-intel-orca-kto |
|
| Data Volume: | |
| Methodology: | | KTO Fine tuning: Kahneman-Tversky Optimization (KTO) |
|
| Context Length: | |
| Model Architecture: | | Auto-regressive language model with optimized transformer architecture |
|
|
| Safety Evaluation |
| Methodologies: | | Evaluation with adversarial datasets, Red teaming exercises |
|
| Risk Categories: | | CBRNE helpfulness, Child Safety, Cyber attack enablement |
|
| Ethical Considerations: | | Safety testing and tuning should be tailored to specific applications; potential for biased, objectionable, or inaccurate outputs. |
|
|
| Responsible Ai Considerations |
| Accountability: | | Developers are responsible for deploying safeguards for their specific use cases. |
|
| Mitigation Strategies: | | Following Responsible Use Guide; integration of safeguards like Llama Guard 3, Prompt Guard, and Code Shield. |
|
|
| Input Output |
| Input Format: | | ChatML or Alpaca prompt template |
|
| Accepted Modalities: | |
| Output Format: | | Multilingual Text and code |
|
|