| Model Type | 
 | |||||||||||||||
| Additional Notes | 
 | |||||||||||||||
| Supported Languages | 
 | |||||||||||||||
| Training Details | 
 | |||||||||||||||
| Input Output | 
 | 
| LLM Name | Mallam 3B 4096 | 
| Repository ๐ค | https://huggingface.co/mesolitica/mallam-3B-4096 | 
| Model Size | 3b | 
| Required VRAM | 6.1 GB | 
| Updated | 2025-09-23 | 
| Maintainer | mesolitica | 
| Model Type | mistral | 
| Model Files | |
| Supported Languages | ms | 
| Model Architecture | MistralForCausalLM | 
| Context Length | 32768 | 
| Model Max Length | 32768 | 
| Transformers Version | 4.36.0.dev0 | 
| Tokenizer Class | PreTrainedTokenizerFast | 
| Padding Token | <pad> | 
| Vocabulary Size | 32000 | 
| Torch Data Type | bfloat16 | 
| Best Alternatives | Context / RAM | Downloads | Likes | 
|---|---|---|---|
| Ministral 3B Instruct | 128K / 6.7 GB | 15833 | 66 | 
| Mistral 3.3B | 128K / 6.6 GB | 17 | 11 | 
| Mistral 3B Instruct V0.2 | 32K / 11.5 GB | 39 | 4 | 
| Dummy Mistral 3B | 32K / 18.2 GB | 17 | 1 | 
| Minerva 3B Ties 1.0 | 16K / 5.8 GB | 756 | 0 | 
| ...inerva 3B Llama3 Instruct V0.1 | 16K / 5.8 GB | 989 | 0 | 
| Asinello Minerva 3B V0.1 | 16K / 5.8 GB | 785 | 0 | 
| Minerva 3B Base RAG | 16K / 5.8 GB | 4 | 12 | 
| Minerva 3B Instruct V1.0 | 16K / 5.8 GB | 15 | 7 | 
| Minerva 3B Instruct V1.0 | 16K / 5.8 GB | 12 | 7 | 
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐