Llama 2 70B GPTQ by localmodels

 Β»  All LLMs  Β»  localmodels  Β»  Llama 2 70B GPTQ   URL Share it on

  4bit   Autotrain compatible   Endpoints compatible   Gptq   Llama   Quantized   Region:us

Llama 2 70B GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 2 70B GPTQ (localmodels/Llama-2-70B-GPTQ)
🌟 Advertise your project πŸš€

Llama 2 70B GPTQ Parameters and Internals

Model Type 
Generative Text
Use Cases 
Areas:
Commercial, Research
Applications:
Assistant-like chat, Natural language generation tasks
Primary Use Cases:
Dialogue, Generating text
Limitations:
Model may produce inaccurate or biased responses
Considerations:
Perform safety and tuning specific to the application before deployment.
Supported Languages 
English (proficient)
Training Details 
Data Sources:
A new mix of publicly available online data
Data Volume:
2 trillion tokens
Methodology:
Supervised fine-tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF)
Context Length:
4096
Hardware Used:
A100-80GB GPUs
Model Architecture:
Optimized transformer architecture
Safety Evaluation 
Methodologies:
Internal evaluations library
Findings:
Outperforms open-source chat models, On par with some closed-source models like ChatGPT and PaLM
Risk Categories:
Misinformation, Bias
Ethical Considerations:
Testing conducted in English; potential biases in non-covered scenarios.
Responsible Ai Considerations 
Mitigation Strategies:
Meta’s sustainability program offsets 100% emissions.
Input Output 
Input Format:
Text only
Accepted Modalities:
Text
Output Format:
Text only
Performance Tips:
Strip() inputs to avoid double-spaces.
LLM NameLlama 2 70B GPTQ
Repository πŸ€—https://huggingface.co/localmodels/Llama-2-70B-GPTQ 
Model Size70b
Required VRAM35.3 GB
Updated2025-09-23
Maintainerlocalmodels
Model Typellama
Model Files  35.3 GB
GPTQ QuantizationYes
Quantization Typegptq|4bit
Model ArchitectureLlamaForCausalLM
Context Length2048
Model Max Length2048
Transformers Version4.32.0.dev0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typefloat16

Best Alternatives to Llama 2 70B GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
...B Instruct AutoRound GPTQ 4bit128K / 39.9 GB12716
...B Instruct AutoRound GPTQ 4bit128K / 39.9 GB10590
...ama 3.1 70B Instruct Gptq 4bit128K / 39.9 GB244
Opus V1.2 70B Marlin32K / 36.4 GB50
MoMo 70B Lora 1.8.4 DPO GPTQ32K / 41.3 GB81
MoMo 70B Lora 1.8.6 DPO GPTQ32K / 41.3 GB51
Midnight Miqu 70B V1.5 GPTQ32G31K / 40.7 GB1894
Tess 70B V1.6 Marlin31K / 36.3 GB71
...Midnight Miqu 70B V1.0 GPTQ32G31K / 40.7 GB72
Senku 70B GPTQ 4bit31K / 36.7 GB61
Note: green Score (e.g. "73.2") means that the model is better than localmodels/Llama-2-70B-GPTQ.

Rank the Llama 2 70B GPTQ Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51534 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124