| Model Type | |
| Use Cases |
| Areas: | |
| Applications: | | General AI systems, Latency and memory-constrained environments |
|
| Primary Use Cases: | | Code, math, and logical reasoning |
|
| Limitations: | | Potential inaccuracy, bias, and harm in high-risk scenarios |
|
| Considerations: | | Models are not evaluated for all downstream use cases, especially high-risk ones. |
|
|
| Additional Notes | | Supports cross-platform capabilities through ONNX runtime across various devices. |
|
| Supported Languages | | primary (English), additional (Multilingual (10% of training data)) |
|
| Training Details |
| Data Sources: | | Publicly available documents, Human-like synthetic data |
|
| Data Volume: | |
| Methodology: | | Supervised fine-tuning and Direct Preference Optimization |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Dense decoder-only Transformer model |
|
|
| Safety Evaluation |
| Methodologies: | | Supervised fine-tuning, Direct Preference Optimization |
|
| Risk Categories: | | Misinformation, Bias, Offensive content |
|
| Ethical Considerations: | | Models may produce unreliable, biased, or offensive outputs. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Models trained on English may over/under-represent groups, or reinforce demeaning stereotypes. |
|
| Transparency: | | Transparency best practices and accountability need to be applied by developers. |
|
| Accountability: | | Developers should ensure compliance with relevant laws and regulations. |
|
| Mitigation Strategies: | | Responsible AI best practices should be followed including the use of safety classifiers. |
|
|
| Input Output |
| Input Format: | | Chat format prompts using user-assistant dialogue. |
|
| Accepted Modalities: | |
| Output Format: | | Generated text in response to input |
|
| Performance Tips: | | Include a BOS token at conversation start for reliable results. |
|
|