| Model Type | |
| Use Cases |
| Applications: | | General purpose AI systems, Research, Commercial applications |
|
| Primary Use Cases: | | Memory/compute constrained environments, Latency bound scenarios, Strong reasoning: code, math, and logic |
|
| Limitations: | | Models may not be suitable for high-risk scenarios without assessments. |
|
| Considerations: | | Models are not specifically evaluated for all downstream purposes. |
|
|
| Additional Notes | | The Phi-3 Medium models can run on multiple platforms with optimized configurations through ONNX. |
|
| Supported Languages | | primary_language (English), other_languages () |
|
| Training Details |
| Data Sources: | | Publicly available documents, High-quality educational data, Code, Newly created synthetic data, High quality chat format supervised data |
|
| Data Volume: | |
| Methodology: | | Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Dense decoder-only Transformer |
|
|
| Responsible Ai Considerations |
| Fairness: | | Models trained primarily on English text which affects other languages and dialects. |
|
| Transparency: | | Developers are advised to follow transparency best practices. |
|
| Accountability: | | Developers are responsible for ensuring compliance with relevant laws and regulations. |
|
| Mitigation Strategies: | | Suggestions for using safety classifiers or custom solutions. |
|
|
| Input Output |
| Input Format: | | Chat format with user and assistant roles. |
|
| Accepted Modalities: | |
| Output Format: | |
|