Phi 2 DPO GGUF by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Phi 2 DPO GGUF   URL Share it on

  Base model:lxuechen/phi-2-dpo Base model:quantized:lxuechen/...   Code   En   Gguf   Model-index   Phi-msft   Quantized   Region:us
Model Card on HF ๐Ÿค—: https://huggingface.co/TheBloke/phi-2-dpo-GGUF 

Phi 2 DPO GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Phi 2 DPO GGUF (TheBloke/phi-2-dpo-GGUF)
๐ŸŒŸ Advertise your project ๐Ÿš€

Phi 2 DPO GGUF Parameters and Internals

Model Type 
text-generation
Use Cases 
Areas:
research
Limitations:
Generate Inaccurate Code and Facts, Limited Scope for code, Unreliable Responses to Instruction, Language Limitations, Potential Societal Biases, Toxicity, Verbosity
Additional Notes 
It was fine-tuned using DPO on a subset of UltraFeedback.
Supported Languages 
english (standard)
Training Details 
Data Sources:
UltraFeedback dataset
Data Volume:
10k subset of UltraFeedback
Methodology:
Direct preference optimization (DPO)
Context Length:
1024
Model Architecture:
earlier SFT model phi-2-sft
Input Output 
Input Format:
### Human: {prompt} ### Assistant:
Accepted Modalities:
text
LLM NamePhi 2 DPO GGUF
Repository ๐Ÿค—https://huggingface.co/TheBloke/phi-2-dpo-GGUF 
Model NamePhi 2 DPO
Model CreatorXuechen Li
Base Model(s)  Phi 2 DPO   lxuechen/phi-2-dpo
Required VRAM1.2 GB
Updated2026-03-12
MaintainerTheBloke
Model Typephi-msft
Model Files  1.2 GB   1.6 GB   1.5 GB   1.2 GB   1.6 GB   1.8 GB   1.6 GB   1.9 GB   2.1 GB   1.9 GB   2.3 GB   3.0 GB
Supported Languagesen
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureAutoModel
Licenseother

Best Alternatives to Phi 2 DPO GGUF

Best Alternatives
Context / RAM
Downloads
Likes
ComicBot V.2 Gguf32K / 5 GB370
Qwen3 Medical GRPO GGUF0K / 1.7 GB1996
...ixtral 8x7B Instruct V0.1 GGUF0K / 15.6 GB25473639
Gemma2 WizardLM0K / 5.2 GB100
Phi 2 GGUF0K / 1.2 GB44293233
Marco O1 GGUF0K / 3 GB2286
Mixtral 8x7B V0.1 GGUF0K / 15.6 GB7823430
Dolphin 2.5 Mixtral 8x7b GGUF0K / 15.6 GB12299303
Dolphin 2.7 Mixtral 8x7b GGUF0K / 15.6 GB9237147
Futfut By Zephyr7b Gguf0K / 5.1 GB14282
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/phi-2-dpo-GGUF.

Rank the Phi 2 DPO GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51629 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124