Phi 3 Mini 128K Instruct is an open-source language model by microsoft. Features: 3.8b LLM, VRAM: 7.7GB, Context: 128K, License: mit, Instruction-Based, HF Score: 68.1, LLM Explorer Score: 0.35, Arc: 63.1, HellaSwag: 80.1, MMLU: 68.7, TruthfulQA: 54.1, WinoGrande: 72.9, GSM8K: 69.5.
Phi 3 Mini 128K Instruct Benchmarks
nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Phi 3 Mini 128K Instruct Parameters and Internals
Model Type
Use Cases
Areas: Research, Commercial applications
Primary Use Cases: Memory/compute constrained environments, Latency bound scenarios, Strong reasoning tasks
Limitations: Not specifically designed or evaluated for all downstream purposes.
Considerations: Adherence to laws and regulations is required.
Additional Notes This is a static model trained on an offline dataset with a cutoff date of October 2023. Future versions may improve upon it.
Supported Languages
Training Details
Data Sources: Publicly available documents, Newly created synthetic data, High quality chat format supervised data
Data Volume:
Methodology: Supervised fine-tuning, Direct Preference Optimization
Context Length:
Training Time:
Hardware Used:
Model Architecture: Dense decoder-only Transformer
Responsible Ai Considerations
Fairness: Models can over- or under-represent groups, erase representation of some groups, or reinforce stereotypes.
Transparency: Inappropriate or offensive content generation potential.
Accountability: Developers need to ensure the model complies with laws and regulations.
Mitigation Strategies: Use safety classifiers or implement custom safety solutions.
Input Output
Input Format:
Accepted Modalities:
Output Format: Generated text in response to input
Performance Tips: For certain GPUs, call AutoModelForCausalLM.from_pretrained() with attn_implementation="eager".
Release Notes
Version:
Notes: Improvement in long-context understanding, instruction following, reasoning capability.
Quantized Models of the Phi 3 Mini 128K Instruct
Best Alternatives to Phi 3 Mini 128K Instruct
Note: green Score (e.g. "73.2 ") means that the model is better than microsoft/Phi-3-mini-128k-instruct .
Expand
Rank the Phi 3 Mini 128K Instruct Capabilities
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
Expand
Check out
Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a