| Model Type | | text generation, instruction tuned |
|
| Use Cases |
| Areas: | |
| Applications: | | instruction-tuned for chat applications |
|
| Primary Use Cases: | | chat-oriented generative tasks |
|
| Limitations: | | Not suitable for language other than English, restricted under Acceptable Use Policy |
|
| Considerations: | | Developers encouraged to implement safety assessments for specific applications. |
|
|
| Additional Notes | | Model supports fine-tuning for languages beyond English under compliance with the license and use policy. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | publicly available online data |
|
| Data Volume: | |
| Methodology: | | supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
| Context Length: | |
| Training Time: | | 7.7M GPU hours on H100-80GB |
|
| Hardware Used: | |
| Model Architecture: | | optimized transformer architecture |
|
|
| Safety Evaluation |
| Methodologies: | | red teaming, adversarial evaluations |
|
| Findings: | | residual risks remain, model refusals reduced |
|
| Risk Categories: | | CBRNE, cybersecurity, child safety |
|
| Ethical Considerations: | | Ethical considerations include avoiding misuse of AI in harmful areas. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Model is optimized to balance helpfulness and alignment, with considerations for avoiding biases. |
|
| Transparency: | | Open source release with detailed documentation and responsible use guidelines. |
|
| Accountability: | | Users must comply with license terms and acceptable use policy. |
|
| Mitigation Strategies: | | Safety tools like Meta Llama Guard and Code Shield provided. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use appropriate hardware and fine-tuning methods for optimal performance. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Initial release of Llama 3 models. |
|
|
|