| Model Type |  | 
| Additional Notes | | Significant improvements in knowledge, coding, mathematics, and instruction following abilities. Multilingual support for more than 29 languages. Long-context support up to 128K tokens, generating up to 8K tokens. Robust against diverse system prompts, enhancing chatbot capabilities. | 
 | 
| Supported Languages | | English (Fluent), Chinese (Fluent), French (Fluent), Spanish (Fluent), Portuguese (Fluent), German (Fluent), Italian (Fluent), Russian (Fluent), Japanese (Fluent), Korean (Fluent), Vietnamese (Fluent), Thai (Fluent), Arabic (Fluent) | 
 | 
| Training Details | 
| Context Length: |  |  | Model Architecture: | | Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias and tied word embeddings | 
 |  | 
| Input Output | 
| Input Format: | | Structured input using predefined roles in JSON format | 
 |  | Accepted Modalities: |  |  | Output Format: |  |  |