Opt 1.3B by facebook

 ยป  All LLMs  ยป  facebook  ยป  Opt 1.3B   URL Share it on

  Arxiv:2005.14165   Arxiv:2205.01068   Autotrain compatible   En   Jax   Opt   Pytorch   Region:us   Tf
Model Card on HF ๐Ÿค—: https://huggingface.co/facebook/opt-1.3b 

Opt 1.3B Benchmarks

Opt 1.3B (facebook/opt-1.3b)
๐ŸŒŸ Advertise your project ๐Ÿš€

Opt 1.3B Parameters and Internals

Model Type 
decoder-only, causal language model
Use Cases 
Areas:
research, evaluation of downstream tasks, text generation
Primary Use Cases:
text generation, evaluation of downstream tasks, fine-tuning
Limitations:
The model has biases due to unfiltered web data, Quality issues in generation diversity and hallucination
Supported Languages 
English (Predominantly), Non-English (Minimal)
Training Details 
Data Sources:
BookCorpus, CC-Stories, The Pile, Pushshift.io Reddit dataset, CCNewsV2
Data Volume:
180B tokens
Methodology:
Causal language modeling
Context Length:
2048
Training Time:
~33 days
Hardware Used:
992 80GB A100 GPUs
Model Architecture:
Decoder-only pre-trained transformer
Input Output 
Input Format:
Sequences of 2048 tokens
Accepted Modalities:
text
Output Format:
Text generation
LLM NameOpt 1.3B
Repository ๐Ÿค—https://huggingface.co/facebook/opt-1.3b 
Model Size1.3b
Required VRAM2.6 GB
Updated2025-08-18
Maintainerfacebook
Model Typeopt
Model Files  2.6 GB
Supported Languagesen
Model ArchitectureOPTForCausalLM
Licenseother
Context Length2048
Model Max Length2048
Transformers Version4.21.0.dev0
Beginning of Sentence Token</s>
End of Sentence Token</s>
Unk Token</s>
Vocabulary Size50272
Torch Data Typefloat16
Activation Functionrelu
Errorsreplace

Quantized Models of the Opt 1.3B

Model
Likes
Downloads
VRAM
Opt 1.3B GPTQ 4bit G1280101 GB
Opt 1.3B Gptq 4bit071 GB

Best Alternatives to Opt 1.3B

Best Alternatives
Context / RAM
Downloads
Likes
LLmRa 1.3B V22K / 5.3 GB6640
... Distilled PromptKD Dolly V1.02K / 2.6 GB50
New Galactica 1.3b Mcq2K / 2.6 GB50
... Galactica 1.3b DPO 0.1beta Ai2K / 2.6 GB31
New Galactica 1.3b Mcq No Rag2K / 2.6 GB60
New Galactica 1.3b Mcq Rag2K / 2.6 GB50
... Galactica 1.3b DPO 0.1beta Ai2K / 2.6 GB50
...tica 1.3b 0.1beta Ai Raft Top32K / 2.6 GB50
...tica 1.3b 0.1beta Ai Raft Top42K / 2.6 GB50
... Galactica 1.3b DPO 0.1beta Ai2K / 2.6 GB50
Note: green Score (e.g. "73.2") means that the model is better than facebook/opt-1.3b.

Rank the Opt 1.3B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50729 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124