SmolLM2 135M Instruct by HuggingFaceTB

 ยป  All LLMs  ยป  HuggingFaceTB  ยป  SmolLM2 135M Instruct   URL Share it on

SmolLM2 135M Instruct is an open-source language model by HuggingFaceTB. Features: 135m LLM, VRAM: 0.3GB, Context: 8K, License: apache-2.0, Instruction-Based, LLM Explorer Score: 0.3.

  Arxiv:2502.02737 Base model:huggingfacetb/smoll... Base model:quantized:huggingfa...   Conversational   En   Endpoints compatible   Instruct   Llama   Onnx   Region:us   Safetensors   Tensorboard   Transformers.js

SmolLM2 135M Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
SmolLM2 135M Instruct (HuggingFaceTB/SmolLM2-135M-Instruct)
๐ŸŒŸ Advertise your project ๐Ÿš€

SmolLM2 135M Instruct Parameters and Internals

Model Type 
language model, instruction following
Use Cases 
Areas:
assistive tools
Applications:
text rewriting, summarization, function calling
Primary Use Cases:
Assistive tools for generating English text
Limitations:
Models generate English content, Factually inaccuracy possible, Logical inconsistencies and biases
Considerations:
Verify important information and critically evaluate.
Additional Notes 
SmolLM2 models offer lightweight capabilities for on-device use.
Supported Languages 
en (primary)
Training Details 
Data Sources:
FineWeb-Edu, DCLM, The Stack, UltraFeedback, Synth-APIGen-v0.1
Data Volume:
2T tokens
Methodology:
Supervised fine-tuning (SFT), Direct Preference Optimization (DPO)
Hardware Used:
64 H100 GPUs
Model Architecture:
Transformer decoder
Input Output 
Accepted Modalities:
text
LLM NameSmolLM2 135M Instruct
Repository ๐Ÿค—https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct 
Base Model(s)  SmolLM2 135M   HuggingFaceTB/SmolLM2-135M
Model Size135m
Required VRAM0.3 GB
Updated2026-04-11
MaintainerHuggingFaceTB
Model Typellama
Instruction-BasedYes
Model Files  0.3 GB   0.0 GB
Supported Languagesen
Context Length8k
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length8192
Model Max Length8192
Transformers Version4.42.3
Tokenizer ClassGPT2Tokenizer
Padding Token<|im_end|>
Vocabulary Size49152
Torch Data Typebfloat16

Quantized Models of the SmolLM2 135M Instruct

Model
Likes
Downloads
VRAM
SmolLM2 135M Instruct Q8 Mlx2670 GB

Best Alternatives to SmolLM2 135M Instruct

Best Alternatives
Context / RAM
Downloads
Likes
SmolLM2 Rethink 135M8K / 0.5 GB111
SmolLM2 135M Instruct Ita8K / 0.1 GB240
...molLM2 135M Instruct Reasoning8K / 0.3 GB70
...wre324 R1 SmolLM2 135M Distill8K / 0.5 GB60
SmolLM2 135M Instruct8K / 0.3 GB22510
Kurtis SmolLM2 135M Instruct8K / 0.5 GB60
...rtis SmolLM2 135M Instruct DPO8K / 0.5 GB50
SmolLM2 135M Grpo Gsm8k8K / 0.5 GB208
Jaja Small V48K / 0.5 GB120
SmolLM2 135M Humanized8K / 0.3 GB450

Rank the SmolLM2 135M Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a