| Model Type | |
| Use Cases |
| Areas: | |
| Applications: | | dialogue, natural language generation |
|
| Primary Use Cases: | | assistant-like chat, natural language generation tasks |
|
| Limitations: | | use only in English, avoid inappropriate legal applications |
|
| Considerations: | | Use proper formatting for optimal feature extraction and performance. |
|
|
| Additional Notes | | AWQ models support efficient inference and faster execution. Compatible with vLLM and AutoAWQ. |
|
| Training Details |
| Data Sources: | | publicly available online data |
|
| Data Volume: | |
| Methodology: | | auto-regressive language model with transformer architecture, supervised fine-tuning (SFT), and reinforcement learning with human feedback (RLHF) |
|
| Context Length: | |
| Hardware Used: | | A100-80GB GPUs (TDP of 350-400W) |
|
| Model Architecture: | |
|
| Safety Evaluation |
| Methodologies: | | truthfulQA, Toxigen benchmarks |
|
| Findings: | | Llama-2-Chat models demonstrate high-performance in safety tests |
|
| Risk Categories: | |
| Ethical Considerations: | | Testing conducted to date has been in English and may not cover all scenarios. Potential for inaccurate or objectionable responses exists. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Testing conducted to date has been in English. Model may exhibit bias. |
|
| Transparency: | | Limited model transparency as it is fine-tuned and pretrained with human feedback. |
|
| Accountability: | | Meta is accountable for the model's performance and outputs. |
|
| Mitigation Strategies: | | Developers advised to perform safety testing and tuning tailored to applications. |
|
|
| Input Output |
| Input Format: | | Structured text with prompt template |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use 'INST' for better performance in chat tasks. |
|
|