DeepSeek V4 Flash 2bit DQ is an open-source language model by mlx-community. Features: 284.3b LLM, VRAM: 96.3GB, Context: 1024K, Quantized, LLM Explorer Score: 0.44.
| LLM Name | DeepSeek V4 Flash 2bit DQ |
| Repository 🤗 | https://huggingface.co/mlx-community/DeepSeek-V4-Flash-2bit-DQ |
| Model Size | 284.3b |
| Required VRAM | 96.3 GB |
| Updated | 2026-04-28 |
| Maintainer | mlx-community |
| Model Type | deepseek_v4 |
| Model Files | |
| Supported Languages | en |
| Quantization Type | 2bit |
| Model Architecture | DeepseekV4ForCausalLM |
| Context Length | 1048576 |
| Model Max Length | 1048576 |
| Transformers Version | 4.57.1 |
| Tokenizer Class | TokenizersBackend |
| Padding Token | <|end▁of▁sentence|> |
| Vocabulary Size | 129280 |
| Torch Data Type | bfloat16 |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
|---|---|---|---|
| DeepSeek V4 Flash 4bit | 1024K / 152.2 GB | 4410 | 4 |
| ...seek Ai DeepSeek V4 Flash 8bit | 1024K / 215.2 GB | 5857 | 10 |
| ...seek Ai DeepSeek V4 Flash 4bit | 1024K / 159.7 GB | 3682 | 2 |
| DeepSeek V4 Flash Mxfp8 | 1024K / 154.8 GB | 1551 | 2 |
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟