Llama 3 8B Instruct Gradient 1048K 8.0bpw H8 EXL2 is an open-source language model by LoneStriker. Features: 8b LLM, VRAM: 8.6GB, Context: 1024K, License: llama3, Quantized, Instruction-Based, LLM Explorer Score: 0.13.
Llama 3 8B Instruct Gradient 1048K 8.0bpw H8 EXL2 Benchmarks
nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 3 8B Instruct Gradient 1048K 8.0bpw H8 EXL2 Parameters and Internals
Model Type
Use Cases
Areas:
Applications:
Primary Use Cases: Instruction tuned models for assistant-like tasks
Limitations: Use outside English laid out by Acceptable Use Policy
Considerations: Developers can fine-tune models for non-English languages adhering to license policy
Supported Languages
Training Details
Data Sources:
Data Volume: 15 trillion tokens of pretraining data
Methodology: Progressive training on increasing context lengths, NTK-aware interpolation to initialize RoPE theta
Context Length:
Training Time:
Hardware Used: Crusoe Energy high performance L40S cluster (GPU), Meta's Research SuperCluster (H100-80GB GPUs)
Model Architecture: Optimized transformer architecture using NTK-aware interpolation and RoPE theta optimization
Safety Evaluation
Methodologies: Red teaming, Adversarial tests
Findings: Residual risks are minimized, focus on limiting false refusals and maintaining model helpfulness
Risk Categories: Cybersecurity risks, Child safety risks, CBRNE hazards
Ethical Considerations: Transparency, rapid feedback loops, community collaboration for safety
Responsible Ai Considerations
Fairness: Model designed to be helpful and unbiased across different use cases
Transparency: Open approach with community feedback to ensure improvements in safety and efficiency
Accountability: Meta ensures accountability through detailed Responsible Use Guide and community interactions
Mitigation Strategies: Deployment of Meta Llama Guard 2 and Code Shield safeguards
Input Output
Input Format:
Accepted Modalities:
Output Format:
Performance Tips: Optimize inputs for long context handling utilizing model's capability
Release Notes
Version:
Date:
Notes: Extended context, improved training efficiency with long contexts
Best Alternatives to Llama 3 8B Instruct Gradient 1048K 8.0bpw H8 EXL2
Note: green Score (e.g. "73.2 ") means that the model is better than LoneStriker/Llama-3-8B-Instruct-Gradient-1048k-8.0bpw-h8-exl2 .
Expand
Rank the Llama 3 8B Instruct Gradient 1048K 8.0bpw H8 EXL2 Capabilities
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
Expand
Check out
Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a