Codellama 34B Bnb 4bit by unsloth

 »  All LLMs  »  unsloth  »  Codellama 34B Bnb 4bit   URL Share it on

Codellama 34B Bnb 4bit is an open-source language model by unsloth. Features: 34b LLM, VRAM: 18.2GB, Context: 16K, License: apache-2.0, Quantized, Code Generating, LLM Explorer Score: 0.11.

  4-bit   4bit   Bitsandbytes   Codegen   Codellama   Codellama-34b   En   Endpoints compatible   Llama   Quantized   Region:us   Safetensors   Sharded   Tensorflow   Unsloth

Codellama 34B Bnb 4bit Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

Codellama 34B Bnb 4bit Parameters and Internals

Model Type 
text generation, transformers
Use Cases 
Areas:
research, commercial applications
Applications:
finetuning models for specific domains
Considerations:
Unsloth provides Colab notebooks for accessible finetuning.
Additional Notes 
Focused on providing convenient, memory-efficient approaches with Colab-supported notebooks for popular models.
Training Details 
Hardware Used:
A100
Input Output 
Accepted Modalities:
text
Performance Tips:
Ensure to use correct resource allocation in Colab for optimal performance.
LLM NameCodellama 34B Bnb 4bit
Repository 🤗https://huggingface.co/unsloth/codellama-34b-bnb-4bit 
Model Size34b
Required VRAM18.2 GB
Updated2026-04-23
Maintainerunsloth
Model Typellama
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   5.0 GB: 3-of-4   3.2 GB: 4-of-4
Supported Languagesen
Quantization Type4bit
Generates CodeYes
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length16384
Model Max Length16384
Transformers Version4.44.2
Tokenizer ClassCodeLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Padding Token<unk>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typebfloat16

Quantized Models of the Codellama 34B Bnb 4bit

Model
Likes
Downloads
VRAM
Codellama Extraction0567 GB

Best Alternatives to Codellama 34B Bnb 4bit

Best Alternatives
Context / RAM
Downloads
Likes
CodeLlama 34B Instruct Hf 4bit16K / 19.4 GB622
CodeLlama 34B Instruct Fp1616K / 67.5 GB28196
CodeLlama 34B Python Fp1616K / 67.5 GB137714
Codellama Extraction16K / 67.6 GB50
Phind Codellama 34B V2 EXL216K /  GB916
CodeLlama 34B Fp1616K / 67.5 GB124
XwinCoder 34B 4.0bpw H6 EXL216K / 17.4 GB61
...Codellama 34B V2 Megacode EXL216K /  GB710
...gpt 32K Codellama 34B Instruct32K / 67.5 GB5022
CodeLlama 34B Instruct Hf16K / 67.5 GB21226304
Note: green Score (e.g. "73.2") means that the model is better than unsloth/codellama-34b-bnb-4bit.

Rank the Codellama 34B Bnb 4bit Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 53232 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a