| Model Type | | text generation, instruction-tuned |
|
| Use Cases |
| Areas: | |
| Applications: | |
| Primary Use Cases: | | natural language generation tasks |
|
| Limitations: | | English only; out-of-scope for illegal or prohibited use cases |
|
| Considerations: | | Developers may fine-tune for languages beyond English adhering to specific policies. |
|
|
| Additional Notes | | The model is intended for English-language applications and could be adapted for other languages. |
|
| Supported Languages | | languages_supported (English), proficiency () |
|
| Training Details |
| Data Sources: | |
| Data Volume: | |
| Methodology: | | Supervised fine-tuning, reinforcement learning with human feedback (RLHF) |
|
| Context Length: | |
| Hardware Used: | | Crusoe Energy high performance L40S cluster |
|
| Model Architecture: | | Transformer with improved RoPE theta, NTK-aware interpolation |
|
|
| Safety Evaluation |
| Methodologies: | | red teaming, adversarial evaluations |
|
| Risk Categories: | | misinformation, bias, cybersecurity, child safety |
|
| Ethical Considerations: | | Iterative testing for CBRNE threats |
|
|
| Responsible Ai Considerations |
| Fairness: | | Model is optimized for safety and helpfulness but trade-offs exist. |
|
| Transparency: | | Steps for safety best practices outlined in Responsible Use Guide. |
|
| Accountability: | |
| Mitigation Strategies: | | Meta Llama Guard 2 and Code Shield provided for safety tailored applications |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Further fine-tuned for assistant-like chat ability; extended context length. |
|
|
|