Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model by jaredjoss

 ยป  All LLMs  ยป  jaredjoss  ยป  Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model   URL Share it on

Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model is an open-source language model by jaredjoss. Features: 410m LLM, VRAM: 1.6GB, Context: 2K, License: mit, LLM Explorer Score: 0.14.

  Autotrain compatible Dataset:jaredjoss/jigsaw-long-...   En   Endpoints compatible   Gpt neox   Region:us   Safetensors

Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model (jaredjoss/pythia-410m-roberta-lr_8e7-kl_01-steps_12000-rlhf-model)
๐ŸŒŸ Advertise your project ๐Ÿš€

Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model Parameters and Internals

Model Type 
text generation
Supported Languages 
en (Unknown)
Training Details 
Data Sources:
jaredjoss/jigsaw-long-2000
Methodology:
Fine-tuning with RLHF
LLM NamePythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model
Repository ๐Ÿค—https://huggingface.co/jaredjoss/pythia-410m-roberta-lr_8e7-kl_01-steps_12000-rlhf-model 
Model Size410m
Required VRAM1.6 GB
Updated2025-09-23
Maintainerjaredjoss
Model Typegpt_neox
Model Files  1.6 GB
Supported Languagesen
Model ArchitectureGPTNeoXForCausalLM
Licensemit
Context Length2048
Model Max Length2048
Transformers Version4.37.0
Tokenizer ClassGPTNeoXTokenizer
Padding Token<|padding|>
Vocabulary Size50304
Torch Data Typefloat32

Best Alternatives to Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model

Best Alternatives
Context / RAM
Downloads
Likes
...thia 410M Cell Type Prediction4K / 0 GB83310
Pythia 410M2K / 0.9 GB18476436
Pythia 410M Sft Full2K / 0.8 GB51
Healix 410M2K / 1.6 GB7820
Pythia 410M Deduped2K / 0.9 GB2681220
Pythia 410M Deduped SimPOW 02K / 0.8 GB50
Pythia 410M Orpo2K / 1.6 GB50
Pythia 410M Ludii Sft2K / 1.6 GB50
... Llm Pythia 410M Pm Gen Ian Nd2K / 1.6 GB50
Pythia410m Sft Tldr2K / 1.6 GB120
Note: green Score (e.g. "73.2") means that the model is better than jaredjoss/pythia-410m-roberta-lr_8e7-kl_01-steps_12000-rlhf-model.

Rank the Pythia 410M Roberta Lr 8e7 Kl 01 Steps 12000 Rlhf Model Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a