Phi 3 Medium 128K Instruct is an open-source language model by microsoft. Features: 14b LLM, VRAM: 28GB, Context: 128K, License: mit, Instruction-Based, HF Score: 73, LLM Explorer Score: 0.28, Arc: 66.5, HellaSwag: 84.9, MMLU: 76.8, TruthfulQA: 54.6, WinoGrande: 74.7, GSM8K: 80.5.
Phi 3 Medium 128K Instruct Benchmarks
nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Phi 3 Medium 128K Instruct Parameters and Internals
Model Type
Use Cases
Areas:
Applications: Language and multimodal models, Generative AI
Primary Use Cases: Memory/compute constrained environments for general AI system, Latency bound scenarios requiring strong reasoning, Research acceleration
Limitations: Not evaluated for all downstream purposes, High risk scenarios may require debiasing techniques
Considerations: Evaluate and mitigate accuracy, safety, and fairness before use in specific downstream scenarios.
Supported Languages
Training Details
Data Sources: Publicly available documents, Newly created synthetic text data, High quality chat format supervised data
Data Volume:
Methodology: Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO)
Context Length:
Training Time:
Hardware Used:
Model Architecture: Dense decoder-only Transformer
Responsible Ai Considerations
Fairness: Trained primarily on English text, under-representation of some English language varieties is possible.
Transparency: Models may generate inappropriate or offensive content; transparent communication with users recommended.
Accountability: Developers should adhere to applicable laws and verify outputs before use in high-risk scenarios.
Mitigation Strategies: Built-in safety classifiers and custom solutions for high-risk contexts.
Input Output
Input Format:
Accepted Modalities:
Output Format:
Performance Tips: Add a BOS token (`~~`) at the start of the conversation for reliable results.
Release Notes
Version:
Date:
Notes: Phi-3-Medium-128K-Instruct model released with 14B parameters and advanced capabilities.
Quantized Models of the Phi 3 Medium 128K Instruct
Best Alternatives to Phi 3 Medium 128K Instruct
Note: green Score (e.g. "73.2 ") means that the model is better than microsoft/Phi-3-medium-128k-instruct .
Expand
Rank the Phi 3 Medium 128K Instruct Capabilities
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
Expand
Check out
Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a