Stablelm 3B 4e1t by stabilityai

 ยป  All LLMs  ยป  stabilityai  ยป  Stablelm 3B 4e1t   URL Share it on

  Arxiv:1607.06450   Arxiv:1910.02054   Arxiv:1910.07467   Arxiv:2101.00027   Arxiv:2104.09864   Arxiv:2204.06745   Arxiv:2305.06161   Arxiv:2307.09288   Autotrain compatible   Dataset:allenai/pes2o   Dataset:bigcode/starcoderdata   Dataset:carperai/pilev2-dev Dataset:tiiuae/falcon-refinedw... Dataset:togethercomputer/redpa...   En   Endpoints compatible   Model-index   Region:us   Safetensors   Stablelm

Stablelm 3B 4e1t Benchmarks

๐ŸŒŸ Advertise your project ๐Ÿš€

Stablelm 3B 4e1t Parameters and Internals

Model Type 
auto-regressive, transformer, decoder-only
Use Cases 
Primary Use Cases:
text generation
Limitations:
May exhibit unreliable, unsafe, or undesirable behaviors requiring correction through evaluation and fine-tuning, Dataset may contain offensive or inappropriate content, Exercise caution for production systems
Considerations:
Not suitable for applications that may cause deliberate or unintentional harm.
Additional Notes 
Recommended to fine-tune the base StableLM-3B-4E1T for downstream tasks.
Supported Languages 
English (proficient)
Training Details 
Data Sources:
tiiuae/falcon-refinedweb, togethercomputer/RedPajama-Data-1T, CarperAI/pilev2-dev, bigcode/starcoderdata, allenai/peS2o
Data Volume:
1 trillion tokens
Methodology:
Pre-trained in bfloat16 precision, optimized with AdamW, trained using the NeoX tokenizer with a vocabulary size of 50,257
Context Length:
4096
Training Time:
30 days
Hardware Used:
256 NVIDIA A100 40GB GPUs (AWS P4d instances)
Model Architecture:
decoder-only transformer, similar to LLaMA, with modifications in position embeddings, normalization, and tokenizer using GPT-NeoX
Responsible Ai Considerations 
Mitigation Strategies:
Users must evaluate and fine-tune the model for safe performance.
LLM NameStablelm 3B 4e1t
Repository ๐Ÿค—https://huggingface.co/stabilityai/stablelm-3b-4e1t 
Model Size3b
Required VRAM5.6 GB
Updated2025-06-09
Maintainerstabilityai
Model Typestablelm
Model Files  5.6 GB
Supported Languagesen
Model ArchitectureStableLmForCausalLM
Licensecc-by-sa-4.0
Context Length4096
Model Max Length4096
Transformers Version4.38.0
Tokenizer ClassGPTNeoXTokenizer
Vocabulary Size50304
Torch Data Typebfloat16
Stablelm 3B 4e1t (stabilityai/stablelm-3b-4e1t)

Quantized Models of the Stablelm 3B 4e1t

Model
Likes
Downloads
VRAM
Stablelm 3B 4e1t GGUF110791 GB

Best Alternatives to Stablelm 3B 4e1t

Best Alternatives
Context / RAM
Downloads
Likes
...t 3B Mix Spider Bird 200 Steps16K / 5.6 GB90
... Instruct 3B Spider 3500 Steps16K / 11.2 GB160
Stablelm Zephyr 3B4K / 5.6 GB15080256
Stablelm 3B Abliterated4K / 5.6 GB210
ReMask 3B4K / 11.2 GB3116
Stablelm 4e1t 2B V0.14K / 4 GB280
Machroom 3B Model Stock4K / 5.6 GB1060
Zephyr Sumbot All Songs4K / 5.6 GB131
Canvers Slm Ov V14K / 2.8 GB130
Ft Stablelm Zephyr 3B4K / 2.5 GB170
Note: green Score (e.g. "73.2") means that the model is better than stabilityai/stablelm-3b-4e1t.

Rank the Stablelm 3B 4e1t Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 48046 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124