WSB GPT 13B by Sentdex

 ยป  All LLMs  ยป  Sentdex  ยป  WSB GPT 13B   URL Share it on

  4-bit   Autotrain compatible   Bitsandbytes Dataset:sentdex/wsb reddit v00...   Endpoints compatible   Llama   Pytorch   Region:us   Sharded
Model Card on HF ๐Ÿค—: https://huggingface.co/Sentdex/WSB-GPT-13B 

WSB GPT 13B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
WSB GPT 13B (Sentdex/WSB-GPT-13B)
๐ŸŒŸ Advertise your project ๐Ÿš€

WSB GPT 13B Parameters and Internals

Model Type 
Instruct, Chat
Use Cases 
Primary Use Cases:
To be a fun chatbot and to learn more about QLoRA.
Limitations:
Prone to using words that were popularly used in the WSB subreddit in that era., Can be offensive and is not intended for all audiences.
Additional Notes 
Despite `~~` being the typical Llama stop token, it was not generated in training/testing, hence โ€˜### END.' was used.
Supported Languages 
multilingual (Multilingual from Llama 2, but not sure what the fine-tune did to it, or if the fine-tuned behavior translates well to other languages.)
Training Details 
Data Sources:
Sentdex/wsb_reddit_v002
Methodology:
Fine-tuned with QLoRA on 2017-2018ish /r/wallstreetbets subreddit comments and responses.
Training Time:
A few hours to train, but days to integrate everything.
Hardware Used:
Lambda Labs 1x H100 80GB GPU instance
Responsible Ai Considerations 
Accountability:
Sentdex
Mitigation Strategies:
As time goes on, some words may be pruned or find-replaced.
Input Output 
Input Format:
### Comment: [parent comment text] ### REPLY: [bot's reply] ### END.
Performance Tips:
Use '### END.' for the model to stop generating, though this isn't ideal.
LLM NameWSB GPT 13B
Repository ๐Ÿค—https://huggingface.co/Sentdex/WSB-GPT-13B 
Model Size13b
Required VRAM26 GB
Updated2025-09-23
MaintainerSentdex
Model Typellama
Model Files  9.9 GB: 1-of-3   9.9 GB: 2-of-3   6.2 GB: 3-of-3
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.33.0.dev0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typebfloat16

Best Alternatives to WSB GPT 13B

Best Alternatives
Context / RAM
Downloads
Likes
Luminaura RP 13B128K / 26 GB60
Yarn Llama 2 13B 128K128K / 26 GB34112
Agent Llama2 13B 80K80K / 26.4 GB50
Chat Llama2 13B 80K80K / 52.8 GB50
LongAlign 13B 64K64K / 26 GB11513
LongAlign 13B 64K Base64K / 26 GB923
LongAlign 13B 64K64K / 26 GB1113
LongAlign 13B 64K Base64K / 26 GB63
Openbuddy Llama2 13B V15p1 64K64K / 26.1 GB64
Openbuddy Llama2 13b64k V1564K / 26.1 GB52
Note: green Score (e.g. "73.2") means that the model is better than Sentdex/WSB-GPT-13B.

Rank the WSB GPT 13B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51534 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124