| Model Type | |
| Use Cases |
| Areas: | | Research, Commercial applications |
|
| Applications: | | General purpose AI systems, Memory/compute constrained environments, Latency bound scenarios |
|
| Primary Use Cases: | | Language and multimodal research, Building blocks for generative AI powered features |
|
| Limitations: | | Not designed for all downstream purposes |
|
| Considerations: | | Developers should consider common limitations and adhere to laws. |
|
|
| Additional Notes | | Optimized for 4K token context length. Quantum int4 ONNX versions available. |
|
| Supported Languages | | language (multilingual), proficiency (primary focus on English) |
|
| Training Details |
| Data Sources: | | Publicly available documents, High-quality educational data, Newly created synthetic data |
|
| Data Volume: | |
| Methodology: | | Supervised fine-tuning and Direct Preference Optimization |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Dense decoder-only Transformer |
|
|
| Safety Evaluation |
| Methodologies: | |
| Ethical Considerations: | | Developers should mitigate accuracy, safety, and fairness risks |
|
|
| Responsible Ai Considerations |
| Fairness: | | The models may reinforce demeaning or negative stereotypes due to training data biases. |
|
| Transparency: | | Developers should follow transparency best practices. |
|
| Accountability: | | Developers are responsible for ensuring compliance with laws and regulations. |
|
| Mitigation Strategies: | | Train with supervision and optimize for safety guidelines. |
|
|
| Input Output |
| Input Format: | | Text format suited for chat prompts. |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Ensure BOS token is included for more reliable results. |
|
|