Nanbeige 16B Chat 32K GPTQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Nanbeige 16B Chat 32K GPTQ   URL Share it on

  4-bit   Autotrain compatible Base model:nanbeige/nanbeige-1... Base model:quantized:nanbeige/...   Custom code   En   Gptq   Nanbeige   Quantized   Region:us   Safetensors   Sharded   Tensorflow   Zh

Nanbeige 16B Chat 32K GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Nanbeige 16B Chat 32K GPTQ (TheBloke/Nanbeige-16B-Chat-32K-GPTQ)
๐ŸŒŸ Advertise your project ๐Ÿš€

Nanbeige 16B Chat 32K GPTQ Parameters and Internals

Model Type 
nanbeige
Use Cases 
Areas:
research, commercial applications
Primary Use Cases:
Q&A systems, text generation
Limitations:
Potential for generating biased or harmful content
Considerations:
Users should ensure outputs comply with ethical and legal standards
Additional Notes 
This model emphasizes safety but acknowledges potential limitations due to probabilistic outputs.
Supported Languages 
en (excellent), zh (excellent)
Training Details 
Data Sources:
internet corpus, books, code
Data Volume:
2.5 trillion tokens for pre-training
Methodology:
Human-aligned training, YaRN interpolation method for position encoding
Context Length:
32000
Input Output 
Input Format:
{prompt}
Accepted Modalities:
text
Output Format:
text
LLM NameNanbeige 16B Chat 32K GPTQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/Nanbeige-16B-Chat-32K-GPTQ 
Model NameNanbeige 16B Chat 32K
Model CreatorNanbeige LLM Lab
Base Model(s)  Nanbeige/Nanbeige-16B-Chat-32K   Nanbeige/Nanbeige-16B-Chat-32K
Model Size16b
Required VRAM9.2 GB
Updated2025-08-31
MaintainerTheBloke
Model Typenanbeige
Model Files  5.0 GB: 1-of-2   4.2 GB: 2-of-2
Supported Languagesen zh
GPTQ QuantizationYes
Quantization Typegptq
Model ArchitectureNanbeigeForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.35.0
Tokenizer ClassNanbeigeTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size59392
Torch Data Typebfloat16

Best Alternatives to Nanbeige 16B Chat 32K GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Nanbeige 16B Base GPTQ4K / 9.2 GB22
Nanbeige 16B Base 32K GPTQ4K / 9.2 GB31
Nanbeige 16B Chat GPTQ4K / 9.2 GB11
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Nanbeige-16B-Chat-32K-GPTQ.

Rank the Nanbeige 16B Chat 32K GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51022 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124