14B is an open-source language model by CausalLM. Features: 14b LLM, VRAM: 28.4GB, Context: 8K, License: wtfpl, Instruction-Based, HF Score: 63.8, LLM Explorer Score: 0.2, Arc: 56.7, HellaSwag: 79.1, MMLU: 65.9, TruthfulQA: 47.8, WinoGrande: 74.9, GSM8K: 58.6.
14B Benchmarks
nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
14B Parameters and Internals
Model Type text-generation, causallm
Use Cases
Areas: Research, Commercial applications
Primary Use Cases:
Limitations: May produce hallucinations or unreliable outputs, Contains objectionable content, pornography, violence, and offensive language
Considerations: Conduct own safety checks, filter keywords
Additional Notes The model is not recommended for quantization, but rather using smaller size models like 7B.
Supported Languages en (English), zh (Chinese)
Training Details
Data Sources: JosephusCheung/GuanacoDataset, Open-Orca/OpenOrca, stingning/ultrachat, meta-math/MetaMathQA, liuhaotian/LLaVA-Instruct-150K, jondurbin/airoboros-3.1, WizardLM/WizardLM_evol_instruct_V2_196k, RyokoAI/ShareGPT52K, RyokoAI/Fandom23K, milashkaarshif/MoeGirlPedia_wikitext_raw_archive, wikipedia, wiki_lingua, fnlp/moss-003-sft-data, garage-bAInd/Open-Platypus, LDJnr/Puffin, openbmb/llava_zh, BAAI/COIG, TigerResearch/tigerbot-zhihu-zh-10k, liwu/MNBVC, teknium/openhermes
Data Volume:
Methodology: Manual or synthetic rewrites, augmented text training, synthetic Wikipedia conversation dataset
Model Architecture: Identical to LLaMA2, using same attention calculation method
Input Output
Input Format: [chatml](https://github.com/openai/openai-python/blob/main/chatml.md)
Accepted Modalities:
Output Format:
Performance Tips: If VRAM is insufficient, use the 7B model instead of the quantized version.
Release Notes
Version:
Date:
Notes: Outperforms Zephyr-ฮฒ in MT-Bench.
Quantized Models of the 14B
Best Alternatives to 14B
Note: green Score (e.g. "73.2 ") means that the model is better than CausalLM/14B .
Expand
Rank the 14B Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
Expand
Check out
Ag3ntum โ our secure, self-hosted AI agent for server management.
Release v20260328a