| LLM Name | DeepSeek V3 1B Test |
| Repository 🤗 | https://huggingface.co/v2ray/DeepSeek-V3-1B-Test |
| Base Model(s) | |
| Model Size | 1b |
| Required VRAM | 2.1 GB |
| Updated | 2025-09-23 |
| Maintainer | v2ray |
| Model Type | deepseek_v3 |
| Model Files | |
| Model Architecture | DeepseekV3ForCausalLM |
| License | mit |
| Context Length | 163840 |
| Model Max Length | 163840 |
| Transformers Version | 4.47.1 |
| Tokenizer Class | LlamaTokenizer |
| Padding Token | <|end▁of▁sentence|> |
| Vocabulary Size | 129280 |
| Torch Data Type | bfloat16 |
🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟