| Model Type | |
| Use Cases |
| Areas: | |
| Applications: | | General purpose AI systems, applications requiring strong reasoning |
|
| Primary Use Cases: | | Memory/compute constrained environments, Latency bound scenarios, Reasoning (code, math, logic) |
|
| Limitations: | | Not evaluated for all downstream purposes, consider AI limitations., Accurate, safe, and fair use in high-risk scenarios require additional evaluations. |
|
| Considerations: | | Adhere to laws and regulations; implement debiasing techniques in applications. |
|
|
| Supported Languages | | Multilingual (English (primary language), other languages (worse performance)) |
|
| Training Details |
| Data Sources: | | Publicly available documents, Filtered documents, High-quality educational data, Code, Synthetic data, Textbook-like data |
|
| Data Volume: | |
| Methodology: | | Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Dense decoder-only Transformer model |
|
|
| Responsible Ai Considerations |
| Fairness: | | These models can over- or under-represent groups or reinforce demeaning stereotypes. |
|
| Transparency: | | Phi series models might be unreliable or offensive. |
|
| Mitigation Strategies: | | Developers should apply debiasing techniques and evaluate for fairness, safety, and accuracy. |
|
|
| Input Output |
| Input Format: | | Prompts using chat format with given templates |
|
| Accepted Modalities: | |
| Output Format: | |
|