| Model Type | |
| Additional Notes | | Significant improvements in knowledge, coding, mathematics, and instruction following abilities. Multilingual support for more than 29 languages. Long-context support up to 128K tokens, generating up to 8K tokens. Robust against diverse system prompts, enhancing chatbot capabilities. |
|
| Supported Languages | | English (Fluent), Chinese (Fluent), French (Fluent), Spanish (Fluent), Portuguese (Fluent), German (Fluent), Italian (Fluent), Russian (Fluent), Japanese (Fluent), Korean (Fluent), Vietnamese (Fluent), Thai (Fluent), Arabic (Fluent) |
|
| Training Details |
| Context Length: | |
| Model Architecture: | | Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias and tied word embeddings |
|
|
| Input Output |
| Input Format: | | Structured input using predefined roles in JSON format |
|
| Accepted Modalities: | |
| Output Format: | |
|