Phi 2 DPO GGUF by TheBloke

 »  All LLMs  »  TheBloke  »  Phi 2 DPO GGUF   URL Share it on

Phi 2 DPO GGUF is an open-source language model by TheBloke. Features: LLM, VRAM: 1.2GB, License: other, Quantized, LLM Explorer Score: 0.12.

  Base model:lxuechen/phi-2-dpo Base model:quantized:lxuechen/...   Code   En   Gguf   Model-index   Phi-msft   Quantized   Region:us

Phi 2 DPO GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

Phi 2 DPO GGUF Parameters and Internals

Model Type 
text-generation
Use Cases 
Areas:
research
Limitations:
Generate Inaccurate Code and Facts, Limited Scope for code, Unreliable Responses to Instruction, Language Limitations, Potential Societal Biases, Toxicity, Verbosity
Additional Notes 
It was fine-tuned using DPO on a subset of UltraFeedback.
Supported Languages 
english (standard)
Training Details 
Data Sources:
UltraFeedback dataset
Data Volume:
10k subset of UltraFeedback
Methodology:
Direct preference optimization (DPO)
Context Length:
1024
Model Architecture:
earlier SFT model phi-2-sft
Input Output 
Input Format:
### Human: {prompt} ### Assistant:
Accepted Modalities:
text
LLM NamePhi 2 DPO GGUF
Repository 🤗https://huggingface.co/TheBloke/phi-2-dpo-GGUF 
Model NamePhi 2 DPO
Model CreatorXuechen Li
Base Model(s)  Phi 2 DPO   lxuechen/phi-2-dpo
Required VRAM1.2 GB
Updated2026-04-19
MaintainerTheBloke
Model Typephi-msft
Model Files  1.2 GB   1.6 GB   1.5 GB   1.2 GB   1.6 GB   1.8 GB   1.6 GB   1.9 GB   2.1 GB   1.9 GB   2.3 GB   3.0 GB
Supported Languagesen
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureAutoModel
Licenseother

Best Alternatives to Phi 2 DPO GGUF

Best Alternatives
Context / RAM
Downloads
Likes
ComicBot V.2 Gguf32K / 5 GB480
Qwen3 Medical GRPO GGUF0K / 1.7 GB1076
Gemma2 WizardLM0K / 5.2 GB100
...ixtral 8x7B Instruct V0.1 GGUF0K / 15.6 GB20684655
Marco O1 GGUF0K / 3 GB956
Phi 2 GGUF0K / 1.2 GB20535232
Dolphin 2.5 Mixtral 8x7b GGUF0K / 15.6 GB15978308
Futfut By Zephyr7b Gguf0K / 5.1 GB23972
Dolphin 2.7 Mixtral 8x7b GGUF0K / 15.6 GB9237147
Mixtral 8x7B V0.1 GGUF0K / 15.6 GB5842437
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/phi-2-dpo-GGUF.

Rank the Phi 2 DPO GGUF Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 53254 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a