LLaMA 33B HF by alexl83

 ยป  All LLMs  ยป  alexl83  ยป  LLaMA 33B HF   URL Share it on

  Autotrain compatible   En   Endpoints compatible   Llama   Pytorch   Region:us   Sharded
Model Card on HF ๐Ÿค—: https://huggingface.co/alexl83/LLaMA-33B-HF 

LLaMA 33B HF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
LLaMA 33B HF (alexl83/LLaMA-33B-HF)
๐ŸŒŸ Advertise your project ๐Ÿš€

LLaMA 33B HF Parameters and Internals

Model Type 
auto-regressive language model
Use Cases 
Areas:
research
Applications:
large language models exploration
Primary Use Cases:
question answering, natural language understanding, reading comprehension
Limitations:
not trained with human feedback, may generate toxic or offensive content
Supported Languages 
en (high proficiency), bg (low proficiency), ca (low proficiency), cs (low proficiency), da (low proficiency), de (low proficiency), es (low proficiency), fr (low proficiency), hr (low proficiency), hu (low proficiency), it (low proficiency), nl (low proficiency), pl (low proficiency), pt (low proficiency), ro (low proficiency), ru (low proficiency), sl (low proficiency), sr (low proficiency), sv (low proficiency), uk (low proficiency)
Training Details 
Data Sources:
CCNet, C4, GitHub, Wikipedia, Books, ArXiv, Stack Exchange
Model Architecture:
transformer
Responsible Ai Considerations 
Fairness:
The model is not intended to inform decisions about matters central to human life. We filtered the data from the Web based on its proximity to Wikipedia text and references.
Mitigation Strategies:
Filtered Web data based on proximity to Wikipedia using a Kneser-Ney language model and a fastText linear classifier.
LLM NameLLaMA 33B HF
Repository ๐Ÿค—https://huggingface.co/alexl83/LLaMA-33B-HF 
Model Size33b
Required VRAM65.4 GB
Updated2025-08-18
Maintaineralexl83
Model Typellama
Model Files  9.8 GB: 1-of-7   10.0 GB: 2-of-7   9.9 GB: 3-of-7   9.9 GB: 4-of-7   9.9 GB: 5-of-7   10.0 GB: 6-of-7   5.9 GB: 7-of-7
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Context Length2048
Model Max Length2048
Transformers Version4.28.0.dev0
Tokenizer ClassLlamaTokenizer
Vocabulary Size32000
Torch Data Typefloat16

Best Alternatives to LLaMA 33B HF

Best Alternatives
Context / RAM
Downloads
Likes
...angled Llama 33M 32K Base V0.132K / 0.1 GB221
ReflectionCoder DS 33B16K / 67 GB100004
Deepseek Wizard 33B Slerp16K / 35.3 GB90
ValidateAI 33B Slerp16K / 35.4 GB50
Deepseek Coder 33B Instruct16K / 66.5 GB16168537
WhiteRabbitNeo 33B V116K / 67 GB173887
Chronos Divergence 33B16K / 65 GB330
ValidateAI 3 33B Ties16K / 66.5 GB60
ValidateAI 2 33B AT16K / 66.5 GB50
...dy Deepseekcoder 33B V16.1 32K16K / 67.1 GB19600
Note: green Score (e.g. "73.2") means that the model is better than alexl83/LLaMA-33B-HF.

Rank the LLaMA 33B HF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50729 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124