Qwen3 MoE Expert Drop Weight Magnitude Pruning R64 S1k 128samples is an open-source language model by jayzou3773. Features: 16b LLM, VRAM: 32.1GB, Context: 40K, MoE.
| LLM Name | Qwen3 MoE Expert Drop Weight Magnitude Pruning R64 S1k 128samples |
| Repository 🤗 | https://huggingface.co/jayzou3773/qwen3-moe-expert_drop-weight_magnitude_pruning-r64-s1k-128samples |
| Model Size | 16b |
| Required VRAM | 32.1 GB |
| Updated | 2026-04-25 |
| Maintainer | jayzou3773 |
| Model Type | qwen3_moe |
| Model Files | |
| Model Architecture | Qwen3MoeForCausalLM |
| Context Length | 40960 |
| Model Max Length | 40960 |
| Transformers Version | 4.53.1 |
| Tokenizer Class | Qwen2Tokenizer |
| Padding Token | <|endoftext|> |
| Vocabulary Size | 151936 |
| Torch Data Type | bfloat16 |
| Errors | replace |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
|---|---|---|---|
| ...wen3 18B A3B Stranger Thoughts | 40K / 34.7 GB | 10 | 0 |
| ... A3B Stranger Thoughts IPONDER | 40K / 34.7 GB | 12 | 0 |
| Qwen3 22B A3B The Harley Quinn | 40K / 44.3 GB | 3 | 1 |
| ...IPONDER Abliterated Uncensored | 40K / 34.7 GB | 8 | 3 |
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟