| Model Type | | text generation, dialogue optimization |
|
| Use Cases |
| Areas: | |
| Applications: | | Assistant-like chat, natural language generation tasks |
|
| Primary Use Cases: | | Assistant applications, Dialogue management |
|
| Limitations: | | Not tested exhaustively across languages, Potential for bias and inaccuracy |
|
| Considerations: | | Developers should ensure thorough safety testing before deployment. |
|
|
| Additional Notes | | This model is static and trained until July 2023. Expected future versions to improve safety based on feedback. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | A new mix of publicly available online data |
|
| Data Volume: | |
| Methodology: | | Includes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
| Context Length: | |
| Hardware Used: | | Meta's Research Super Cluster, A100-80GB GPUs, with cumulative 3.3M GPU hours |
|
| Model Architecture: | | Auto-regressive language model using optimized transformer architecture |
|
|
| Safety Evaluation |
| Methodologies: | |
| Findings: | | Potentially unpredictable outputs, model may produce inaccurate or biased responses |
|
| Risk Categories: | | Inaccuracy, Bias, Objectionable responses |
|
| Ethical Considerations: | | Pre-deployment safety testing recommended |
|
|
| Responsible Ai Considerations |
| Fairness: | | Testing for fairness and bias conducted in English |
|
| Transparency: | | Reports available for potential risks |
|
| Accountability: | | Users are responsible for testing tailored to specific applications |
|
| Mitigation Strategies: | | Recommendations to perform safety tuning tailored to specific applications |
|
|
| Input Output |
| Input Format: | | Text format with specific tags and tokens such as [INST] and <> |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Ensure the correct sequence and format of tokens for the best performance. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Initial release for commercial and research use, focusing on dialogue optimization. |
|
|
|