| Model Type | |
| Use Cases |
| Areas: | | Commercial applications, Research |
|
| Applications: | | Natural language generation tasks |
|
| Primary Use Cases: | |
| Limitations: | | Use in languages other than English, Violation of applicable laws and regulations |
|
| Considerations: | | Specific formatting required for expected performance in chat versions |
|
|
| Additional Notes | | Quantization methods include options like Q2_K, Q3_K, Q4_K, etc., for trade-offs between memory size and model accuracy. |
|
| Supported Languages | | English (unknown proficiency level) |
|
| Training Details |
| Data Sources: | | A new mix of publicly available online data |
|
| Data Volume: | |
| Methodology: | | Pretraining and fine-tuning with supervised fine-tuning and reinforcement learning with human feedback |
|
| Context Length: | |
| Training Time: | | January 2023 to July 2023 |
|
| Hardware Used: | |
| Model Architecture: | | Auto-regressive language model with an optimized transformer architecture |
|
|
| Safety Evaluation |
| Methodologies: | | Internal evaluations library |
|
| Risk Categories: | |
| Ethical Considerations: | | Testing in languages other than English not conducted |
|
|
| Responsible Ai Considerations |
| Fairness: | | Testing conducted for fairness, but not exhaustive |
|
| Transparency: | | Model card available with detailed information |
|
| Accountability: | | Developers accountable for safe deployment of applications |
|
| Mitigation Strategies: | | Use Responsible Use Guide for deployment |
|
|
| Input Output |
| Input Format: | | Text input with special token formatting |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Follow recommended formatting with special tokens for chat models. |
|
|
| Release Notes |
| Version: | | GGUF format introduced on August 21st 2023 |
|
| Date: | |
| Notes: | | New format for improved tokenization and metadata support. |
|
| Version: | | Macro-scaling models with parameter variations (7B, 13B, 70B) |
|
| Notes: | | Pretrained and fine-tuned generative text models available. |
|
|
|