| Model Type | |
| Use Cases |
| Areas: | |
| Applications: | | assistant-like chat, natural language generation tasks |
|
| Primary Use Cases: | | pretrained: general language generation, tuned: chat and assistance |
|
| Limitations: | | Out-of-scope uses that violate policies or laws, limited to English |
|
| Considerations: | | Possible inaccuracies, biases, and objectionable content. |
|
|
| Additional Notes | | Supports long contexts over 1040K. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | |
| Data Volume: | | 1.4B tokens total for all stages |
|
| Methodology: | | NTK-aware interpolation for RoPE theta, progressive training on increasing context lengths |
|
| Context Length: | |
| Hardware Used: | |
| Model Architecture: | | auto-regressive transformer with RingAttention |
|
|
| Safety Evaluation |
| Methodologies: | | extensive red teaming, adversarial evaluations |
|
| Risk Categories: | | CBRNE, Cyber Security, Child Safety |
|
| Ethical Considerations: | | Iterative testing during model training to assess the safety of responses related to CBRNE threats and other adversarial risks. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Safety benchmark standards transparency, comprehensive safety safeguards. |
|
| Transparency: | | Open approach to AI with community involvement. |
|
| Accountability: | | Developers responsible for safety deployment based on use case. |
|
| Mitigation Strategies: | | Use of Purple Llama solutions, thorough safety guides. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use supervised fine-tuning and reinforcement learning with human feedback for optimal results. |
|
|