| LLM Name | TrelisLM 100M Layer Hidden Pruned |
| Repository ๐ค | https://huggingface.co/Trelis/TrelisLM-100M-layer-hidden-pruned |
| Model Size | 100m |
| Required VRAM | 0.2 GB |
| Updated | 2024-09-18 |
| Maintainer | Trelis |
| Model Type | llama |
| Model Files | |
| Model Architecture | LlamaForCausalLM |
| Context Length | 2048 |
| Model Max Length | 2048 |
| Transformers Version | 4.44.2 |
| Tokenizer Class | GPT2Tokenizer |
| Vocabulary Size | 49152 |
| Torch Data Type | bfloat16 |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
|---|---|---|---|
| SmollerLM2 100M Instruct Sft | 8K / 0.3 GB | 11 | 0 |
| SmollerLM2 100M | 8K / 0.2 GB | 5 | 0 |
| Stockmark 2 100B Instruct Beta | 4K / 192.9 GB | 47 | 12 |
| Stockmark 100B | 4K / 191.9 GB | 1459 | 34 |
| Saily 100b | 4K / 235.5 GB | 1891 | 7 |
| Plankton 100M | 4K / 0.4 GB | 7 | 0 |
| Reglu 100B | 2K / 2.6 GB | 1 | 1 |
| ...ephyr Smol Llama 100M DPO Full | 1K / 0.2 GB | 4 | 3 |
| Llama 161M 100B | 1K / 0.3 GB | 15 | 23 |
| ...ephyr Smol Llama 100M DPO Full | 1K / GB | 4 | 1 |
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐