Model Type | |
Use Cases |
Areas: | |
Applications: | assistant-like chat, natural language generation tasks |
|
Primary Use Cases: | dialogue applications, function calling using structured JSON |
|
Limitations: | May produce inaccurate or biased responses, Evaluation primarily in English |
|
Considerations: | Developers should perform safety testing and tuning tailored to specific applications of the model. |
|
|
Additional Notes | GPTQ-trained models offer fast and reliable performance with adapters. |
|
Supported Languages | English (Primary language for training and evaluation.) |
|
Training Details |
Data Sources: | Publicly available online data, Publicly available instruction datasets |
|
Data Volume: | |
Methodology: | fine-tuning using GPTQ and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Training Time: | January 2023 to July 2023 |
|
Hardware Used: | A100-80GB GPUs with TDP of 350-400W |
|
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | Internal safety evaluations, Testing on TruthfulQA and Toxigen benchmarks |
|
Findings: | Llama-2-Chat models produced 0 percent toxic generations in Toxigen evaluations |
|
Risk Categories: | |
Ethical Considerations: | Developers should perform safety testing tailored to specific applications. |
|
|
Responsible Ai Considerations |
Fairness: | |
Transparency: | Model responses might not cover all scenarios in English. |
|
Accountability: | Developers are responsible for safety testing and tuning. |
|
Mitigation Strategies: | Fine-tuning with reinforcement learning and human feedback |
|
|
Input Output |
Input Format: | Structured input with INST and JSON function call formats. |
|
Accepted Modalities: | |
Output Format: | JSON object with function name and arguments. |
|
Performance Tips: | Use GPU for fast and accurate inference. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Includes function calling capabilities. |
|
|
|