| Model Type | |
| Use Cases |
| Areas: | |
| Applications: | |
| Primary Use Cases: | | instruction-tuned models for dialogue |
|
| Limitations: | | Use must comply with laws and Llama 3 license policy |
|
| Considerations: | | Developers should perform safety testing and tuning prior to deployment. |
|
|
| Additional Notes | | Models are optimized for helpfulness and safety through RLHF and SFT. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | publicly available online data |
|
| Data Volume: | |
| Methodology: | | pre-trained, instruction-tuned, RLHF |
|
| Context Length: | |
| Hardware Used: | | Meta's Research SuperCluster, H100-80GB GPUs |
|
| Model Architecture: | | auto-regressive transformer with Grouped-Query Attention |
|
|
| Safety Evaluation |
| Methodologies: | | red teaming, adversarial evaluations |
|
| Findings: | | reduced residual risk via safety mitigations |
|
| Risk Categories: | | misinformation, cybersecurity, child safety |
|
| Ethical Considerations: | | developers must assess risks for specific use cases |
|
|
| Responsible Ai Considerations |
| Fairness: | | Safety benchmarks are transparent and rigorous. |
|
| Transparency: | | Evaluations and benchmarks are publicly accessible. |
|
| Accountability: | | Users and developers must adhere to guidelines and policies. |
|
| Mitigation Strategies: | | Incorporate safeguards like Meta Llama Guard 2 and Code Shield |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use with transformers pipeline or llama3 codebase for best results. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Initial release of Llama 3 models, including optimized transformers architecture. |
|
|
|