We do not advise you to use base language models for text generation. Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., on this model.
Supported Languages
languages_list (Multilingual support of both base and chat models), proficiency_level (High)
Training Details
Methodology:
Transformer architecture with SwiGLU activation, attention QKV bias, group query attention, mixture of sliding window attention and full attention.
Note: green Score (e.g. "73.2") means that the model is better than Qwen/Qwen1.5-32B.
Rank the Qwen1.5 32B Capabilities
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 53254 in total.