| Model Type | | text-generation, causallm |
|
| Use Cases |
| Areas: | | Research, Commercial applications |
|
| Primary Use Cases: | |
| Limitations: | | May produce hallucinations or unreliable outputs, Contains objectionable content, pornography, violence, and offensive language |
|
| Considerations: | | Conduct own safety checks, filter keywords |
|
|
| Additional Notes | | The model is not recommended for quantization, but rather using smaller size models like 7B. |
|
| Supported Languages | | en (English), zh (Chinese) |
|
| Training Details |
| Data Sources: | | JosephusCheung/GuanacoDataset, Open-Orca/OpenOrca, stingning/ultrachat, meta-math/MetaMathQA, liuhaotian/LLaVA-Instruct-150K, jondurbin/airoboros-3.1, WizardLM/WizardLM_evol_instruct_V2_196k, RyokoAI/ShareGPT52K, RyokoAI/Fandom23K, milashkaarshif/MoeGirlPedia_wikitext_raw_archive, wikipedia, wiki_lingua, fnlp/moss-003-sft-data, garage-bAInd/Open-Platypus, LDJnr/Puffin, openbmb/llava_zh, BAAI/COIG, TigerResearch/tigerbot-zhihu-zh-10k, liwu/MNBVC, teknium/openhermes |
|
| Data Volume: | |
| Methodology: | | Manual or synthetic rewrites, augmented text training, synthetic Wikipedia conversation dataset |
|
| Model Architecture: | | Identical to LLaMA2, using same attention calculation method |
|
|
| Input Output |
| Input Format: | | [chatml](https://github.com/openai/openai-python/blob/main/chatml.md) |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | If VRAM is insufficient, use the 7B model instead of the quantized version. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Outperforms Zephyr-ฮฒ in MT-Bench. |
|
|
|