Intended as a compact and efficient model within computational and memory constraints. It is based on optimizations of the Llama architecture for use in respective applications.
Supported Languages
en (proficient)
Training Details
Data Sources:
cerebras/SlimPajama-627B, bigcode/starcoderdata, teknium/openhermes, Dolphin 2.8 dataset by Eric Hartford
Data Volume:
3 trillion tokens
Methodology:
Pretrained using optimized techniques on Llama architecture and tokenizer
Training Time:
90 days
Hardware Used:
16 A100-40G GPUs
Model Architecture:
Same architecture and tokenizer as Llama 2. Compact with only 1.1B parameters.
Note: green Score (e.g. "73.2") means that the model is better than cognitivecomputations/TinyDolphin-2.8-1.1b.
Rank the TinyDolphin 2.8 1.1B Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52392 in total.