Phi 3.5 Mini Instruct by microsoft

 »  All LLMs  »  microsoft  »  Phi 3.5 Mini Instruct   URL Share it on

Phi 3.5 Mini Instruct is an open-source language model by microsoft. Features: 3.8b LLM, VRAM: 7.7GB, Context: 128K, License: mit, Instruction-Based, LLM Explorer Score: 0.36.

  Arxiv:2403.06412   Arxiv:2404.14219   Arxiv:2407.13833   Autotrain compatible   Code   Conversational   Custom code   Endpoints compatible   Instruct   Multilingual   Phi3   Region:us   Safetensors   Sharded   Tensorflow

Phi 3.5 Mini Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

Phi 3.5 Mini Instruct Parameters and Internals

Model Type 
text-generation
Use Cases 
Areas:
research, commercial applications
Applications:
AI systems, natural language processing
Primary Use Cases:
memory/compute constrained environments, latency-bound scenarios, strong reasoning tasks
Limitations:
languages other than English may have worse performance
Considerations:
Developers should consider model limitations and adhere to safety and regulatory guidelines.
Additional Notes 
None
Supported Languages 
Arabic (supported), Chinese (supported), Czech (supported), Danish (supported), Dutch (supported), English (supported), Finnish (supported), French (supported), German (supported), Hebrew (supported), Hungarian (supported), Italian (supported), Japanese (supported), Korean (supported), Norwegian (supported), Polish (supported), Portuguese (supported), Russian (supported), Spanish (supported), Swedish (supported), Thai (supported), Turkish (supported), Ukrainian (supported)
Training Details 
Data Sources:
publicly available documents, textbook-like synthetic data
Data Volume:
3.4T tokens
Methodology:
supervised fine-tuning, proximal policy optimization, and direct preference optimization
Context Length:
128000
Training Time:
10 days
Hardware Used:
512 H100-80G GPUs
Model Architecture:
dense decoder-only Transformer
Safety Evaluation 
Methodologies:
red-teaming, adversarial conversation simulations
Findings:
models may refuse undesirable outputs in English across multiple languages
Risk Categories:
misinformation, bias
Ethical Considerations:
Industry-wide investment in high-quality safety evaluation datasets is needed.
Responsible Ai Considerations 
Fairness:
Models may over- or under-represent groups of people and need fine-tuning for diversity.
Transparency:
Model operation and biases should be understood and communicated to users.
Accountability:
Microsoft accountable for model's outputs.
Mitigation Strategies:
Utilize safety classifiers and fine-tuning based on deployment scenarios.
Input Output 
Input Format:
Text inputs with chat format expected
Accepted Modalities:
text
Output Format:
Generated text
Performance Tips:
Use in-memory or latency-bound scenarios.
Release Notes 
Version:
June 2024
Date:
2024-06
Notes:
Updated with feedback, improved conversation quality in multilingual settings.
LLM NamePhi 3.5 Mini Instruct
Repository 🤗https://huggingface.co/microsoft/Phi-3.5-mini-instruct 
Model Size3.8b
Required VRAM7.7 GB
Updated2025-09-23
Maintainermicrosoft
Model Typephi3
Instruction-BasedYes
Model Files  5.0 GB: 1-of-2   2.7 GB: 2-of-2
Model ArchitecturePhi3ForCausalLM
Licensemit
Context Length131072
Model Max Length131072
Transformers Version4.43.3
Tokenizer ClassLlamaTokenizer
Padding Token<|endoftext|>
Vocabulary Size32064
Torch Data Typebfloat16

Quantized Models of the Phi 3.5 Mini Instruct

Model
Likes
Downloads
VRAM
Phi 3.5 Mini Instruct GGUF263779880 GB
Sign Language 20250823 1904510127 GB
Phi 3.5 Mini Instruct Bnb 4bit13344292 GB
Flow Judge V0.1 AWQ627222 GB

Best Alternatives to Phi 3.5 Mini Instruct

Best Alternatives
Context / RAM
Downloads
Likes
Phi 4 Mini Instruct Heretic128K / 7.7 GB4230
MediPhi Instruct128K / 7.7 GB270966
Phi 3 Mini 128K Instruct128K / 7.7 GB50
NuExtract V1.5128K / 7.7 GB10851189
Phi 3.5 Mini TitanFusion 0.1128K / 7.7 GB180
Phi 4 Mini Instruct128K / 7.7 GB719824
ECE EIFFEL 3Bv2128K / 7.7 GB90
MediPhi Clinical128K / 7.7 GB51214
MediPhi MedCode128K / 7.7 GB3107
MediPhi PubMed128K / 7.7 GB20811

Rank the Phi 3.5 Mini Instruct Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 53185 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a