LLM Name | Gpt2 Large Sft |
Repository ๐ค | https://huggingface.co/gumran/gpt2-large-sft |
Model Name | gpt2-large-sft |
Base Model(s) | |
Model Size | 774m |
Required VRAM | 3.1 GB |
Updated | 2025-08-30 |
Maintainer | gumran |
Model Type | gpt2 |
Model Files | |
Model Architecture | GPT2LMHeadModel |
License | mit |
Model Max Length | 1024 |
Transformers Version | 4.52.4 |
Tokenizer Class | GPT2Tokenizer |
Padding Token | <|im_end|> |
Vocabulary Size | 50259 |
Torch Data Type | float32 |
Activation Function | gelu_new |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
---|---|---|---|
ProtGPT2 With Pad | 0K / 3.1 GB | 1217 | 0 |
Gpt2 Large Sft Imdb | 0K / 1.6 GB | 25 | 0 |
Gpt2 Large Sft Golden Hh | 0K / 1.6 GB | 7 | 0 |
Gpt2 Large Sft Anthropic Hh | 0K / 1.6 GB | 7 | 0 |
Alpaca Refine Gpt2 E1 Se0 | 0K / 3.1 GB | 1081 | 0 |
Alpaca Tuned Gpt2 | 0K / 3.1 GB | 1080 | 0 |
Alpaca Spin Gpt2 E0 Se1 | 0K / 3.1 GB | 1082 | 0 |
Alpaca Spin Gpt2 E1 Se0 | 0K / 3.1 GB | 1066 | 0 |
...T2 774M CINDER SHOW MULTI CHAT | 0K / 0 GB | 2 | 2 |
Alpaca Refine Gpt2 E0 Se1 | 0K / 3.1 GB | 1071 | 0 |
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐