LLaMA3 Iterative DPO Final GGUF by sirovub

 ยป  All LLMs  ยป  sirovub  ยป  LLaMA3 Iterative DPO Final GGUF   URL Share it on

LLaMA3 Iterative DPO Final GGUF is an open-source language model by sirovub. Features: LLM, VRAM: 16.1GB, Context: 8K, License: llama3, Quantized, LLM Explorer Score: 0.14.

  Arxiv:2312.11456   Arxiv:2405.07863   Conversational   Endpoints compatible   Gguf   Llama   Q8   Quantized   Region:us   Sharded   Tensorflow

LLaMA3 Iterative DPO Final GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
LLaMA3 Iterative DPO Final GGUF (sirovub/LLaMA3-iterative-DPO-final-GGUF)
๐ŸŒŸ Advertise your project ๐Ÿš€

LLaMA3 Iterative DPO Final GGUF Parameters and Internals

Model Type 
text generation, instruction-following
Additional Notes 
RLHFlow\LLaMA3-iterative-DPO-final is an unofficial checkpoint developed for research purposes. While safety and ethical considerations are integral to the alignment process, there remains the possibility that the model could generate offensive or unethical content under adversarial conditions.
Training Details 
Data Sources:
https://huggingface.co/datasets/hendrydong/preference_700K, https://huggingface.co/datasets/RLHFlow/prompt-collection-v0.1
Methodology:
Iterative DPO
LLM NameLLaMA3 Iterative DPO Final GGUF
Repository ๐Ÿค—https://huggingface.co/sirovub/LLaMA3-iterative-DPO-final-GGUF 
Required VRAM16.1 GB
Updated2026-04-08
Maintainersirovub
Model Typellama
Model Files  8.5 GB   16.1 GB   5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4
GGUF QuantizationYes
Quantization Typegguf|q8
Model ArchitectureLlamaForCausalLM
Licensellama3
Context Length8192
Model Max Length8192
Transformers Version4.40.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|end_of_text|>
Vocabulary Size128256
Torch Data Typebfloat16

Best Alternatives to LLaMA3 Iterative DPO Final GGUF

Best Alternatives
Context / RAM
Downloads
Likes
KernelLLM GGUF128K / 2.2 GB2312
LLAMA2 GOOD GGUF16K / 4.8 GB170
Codellama Cairo Instruct GGUF16K / 4.1 GB261
Aware Ai 1st8K / 16.1 GB120
MFANNv0.6 GGUF8K / 4.7 GB50
UlizaLlama Q4 K M Gguf4K / 4.2 GB340
Tinyllama Coder Py V164K / 0.7 GB100
Tinyllama Coder Py V154K / 0.7 GB100
Cancer Llama.5 Llm4K / 4.1 GB160
Airavata GGUF4K / 4.2 GB482
Note: green Score (e.g. "73.2") means that the model is better than sirovub/LLaMA3-iterative-DPO-final-GGUF.

Rank the LLaMA3 Iterative DPO Final GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a