Gpt2023 by crumb

 »  All LLMs  »  crumb  »  Gpt2023   URL Share it on

Gpt2023 is an open-source language model by crumb. Features: 137m LLM, VRAM: 0.3GB, License: mit, HF Score: 28.2, LLM Explorer Score: 0.27, ELO: 1345, Arc: 21.9, HellaSwag: 31.1, MMLU: 25.1, TruthfulQA: 40.7, WinoGrande: 50.1, GSM8K: 0.3.

  En   Endpoints compatible   Gpt2   Pytorch   Region:us   Safetensors
Model Card on HF 🤗: https://huggingface.co/crumb/gpt2023 

Gpt2023 Benchmarks

Gpt2023 Parameters and Internals

Model Type 
causal-lm
Use Cases 
Areas:
Research
Limitations:
Lack of awareness of some recent events due to finetuning on a limited dataset
Supported Languages 
en (Fluent)
Training Details 
Data Sources:
common crawl sites, ArXiv, GitHub
Data Volume:
2.23 billion tokens
Methodology:
Finetuning on existing GPT-2 model with learning rate adjustments
Context Length:
1024
Training Time:
79.32 hours
Hardware Used:
12GB RTX3060
Model Architecture:
Transformer-based architecture, left-to-right causal language model
Input Output 
Input Format:
Text input, up to 1024 tokens
Accepted Modalities:
text
Output Format:
Text generation
Performance Tips:
Setting a seed can help achieve reproducible results
LLM NameGpt2023
Repository 🤗https://huggingface.co/crumb/gpt2023 
Model Size137m
Required VRAM0.3 GB
Updated2026-04-12
Maintainercrumb
Model Typegpt2
Model Files  0.3 GB   0.3 GB
Supported Languagesen
Model ArchitectureGPT2LMHeadModel
Licensemit
Model Max Length1024
Transformers Version4.29.0.dev0
Tokenizer ClassGPT2Tokenizer
Vocabulary Size50257
Torch Data Typebfloat16
Activation Functiongelu_new

Best Alternatives to Gpt2023

Best Alternatives
Context / RAM
Downloads
Likes
Gpt20K / 0.5 GB159641843230
GPT2 137M Reasoner V1.00K / 0.5 GB91
Phantasor 137M0K / 0.5 GB2111
Phantasor V0.1 137M0K / 0.5 GB931
Phantasor V0.2 137M0K / 0.5 GB731
Phantasor V0.3 137M0K / 0.5 GB371
Gpt2 Auth0K / 0.5 GB690
My GPT20K / 0.5 GB8090
Gpt2 Test0K / 0.5 GB10170
Xuanxuan0K / 0.3 GB70

Rank the Gpt2023 Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 53570 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a