SmolLM2 1.7B Instruct by HuggingFaceTB

 ยป  All LLMs  ยป  HuggingFaceTB  ยป  SmolLM2 1.7B Instruct   URL Share it on

  Arxiv:2502.02737   Autotrain compatible Base model:huggingfacetb/smoll... Base model:quantized:huggingfa...   Conversational   En   Endpoints compatible   Instruct   Llama   Onnx   Region:us   Safetensors   Tensorboard   Transformers.js

SmolLM2 1.7B Instruct Benchmarks

๐ŸŒŸ Advertise your project ๐Ÿš€

SmolLM2 1.7B Instruct Parameters and Internals

Model Type 
text generation, instruction following
Use Cases 
Areas:
text generation, instruction following, text rewriting, summarization, function calling
Applications:
educational tools, coding assistance, customer support, chatbots, language learning
Primary Use Cases:
text and instruction generation
Limitations:
primarily understands and generates content in English, generated content may not always be factually accurate or free from bias
Considerations:
Users should always verify important information and critically evaluate any generated content.
Supported Languages 
en (main)
Training Details 
Data Sources:
FineWeb-Edu, DCLM, The Stack, new mathematics and coding datasets
Data Volume:
11T tokens
Methodology:
Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO)
Hardware Used:
256 H100 GPUs
Model Architecture:
Transformer decoder
Input Output 
Input Format:
Expected input format includes system and user prompts
Accepted Modalities:
text
Output Format:
text
Performance Tips:
Ensure queries are clear and check model's language support for optimal results.
LLM NameSmolLM2 1.7B Instruct
Repository ๐Ÿค—https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct 
Base Model(s)  HuggingFaceTB/SmolLM2-1.7B   HuggingFaceTB/SmolLM2-1.7B
Model Size1.7b
Required VRAM3.4 GB
Updated2025-06-09
MaintainerHuggingFaceTB
Model Typellama
Instruction-BasedYes
Model Files  3.4 GB   0.0 GB
Supported Languagesen
Context Length8k
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length8192
Model Max Length8192
Transformers Version4.42.3
Tokenizer ClassGPT2Tokenizer
Padding Token<|im_end|>
Vocabulary Size49152
Torch Data Typebfloat16
SmolLM2 1.7B Instruct (HuggingFaceTB/SmolLM2-1.7B-Instruct)

Quantized Models of the SmolLM2 1.7B Instruct

Model
Likes
Downloads
VRAM
SmolLM2 1.7B Instruct Bnb 4bit210481 GB

Best Alternatives to SmolLM2 1.7B Instruct

Best Alternatives
Context / RAM
Downloads
Likes
SmolLM2 1.7B Instruct 16K16K / 3.4 GB169
Penny 1.7B8K / 6.9 GB23723
SmolTulu 1.7B Reinforced8K / 3.4 GB4775
Superthoughts Lite V18K / 3.4 GB203
SmolLM2 1.7 Persona8K / 3.5 GB180
SmolLM2 1.7B Instruct8K / 3.4 GB31924
...ghts Lite 1.8B Experimental O18K / 3.6 GB222
Cllm 1.0.0 340000 Instruct8K / 3.5 GB130
SmolLM2 1.7B Open Thought8K / 3.4 GB270
SmolLM2 1.7B R1 Distilled8K / 3.4 GB120

Rank the SmolLM2 1.7B Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 48046 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124