| Model Type | | chat model, language model |
|
| Additional Notes | | Supports deepspeed-based training and multi-round data construction with enhanced long text generation capabilities. |
|
| Training Details |
| Data Sources: | |
| Data Volume: | | 1.5万亿 Tokens for 7B and 3万亿 Tokens for 12B |
|
| Model Architecture: | |
|
| Input Output |
| Performance Tips: | | Supports multi-round abilities and better focus on multi-round answers. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Release of 12B-v2 version chat model with quantized versions. |
|
| Version: | |
| Date: | |
| Notes: | | Release of 12B version chat model with quantized versions. |
|
| Version: | |
| Date: | |
| Notes: | | Release of 1T Chinese dataset. |
|
| Version: | |
| Date: | |
| Notes: | | Release of 7B version chat model with quantized versions. |
|
|
|