| LLM Name | Meta Llama Llama 4 Scout 17B 16E 4bit |
| Repository ๐ค | https://huggingface.co/mlx-community/meta-llama-Llama-4-Scout-17B-16E-4bit |
| Base Model(s) | |
| Model Size | 17b |
| Required VRAM | 60.5 GB |
| Updated | 2025-10-03 |
| Maintainer | mlx-community |
| Model Type | llama4 |
| Model Files | |
| Supported Languages | ar de en es fr hi id it pt th tl vi |
| Quantization Type | 4bit |
| Model Architecture | Llama4ForConditionalGeneration |
| License | other |
| Model Max Length | 262144 |
| Transformers Version | 4.51.0.dev0 |
| Tokenizer Class | PreTrainedTokenizer |
| Padding Token | <|finetune_right_pad_id|> |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
|---|---|---|---|
| ...Maverick 17B 16E Instruct 4bit | 0K / 146.3 GB | 1394 | 7 |
| ...Maverick 17B 16E Instruct 6bit | 0K / 213.8 GB | 1126 | 2 |
| ...ama Llama 4 Scout 17B 16E Fp16 | 0K / 207.3 GB | 920 | 3 |
| ...ama Llama 4 Scout 17B 16E Bf16 | 0K / 207.3 GB | 1064 | 2 |
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐