Octocoder GPTQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Octocoder GPTQ   URL Share it on

  Arxiv:2308.07124   4-bit   Autotrain compatible   Base model:bigcode/octocoder Base model:quantized:bigcode/o...   Code   Codegen   Dataset:bigcode/commitpackft   Dataset:bigcode/oasst-octopack   Gpt bigcode   Gptq   Model-index   Quantized   Region:us   Safetensors
Model Card on HF ๐Ÿค—: https://huggingface.co/TheBloke/Octocoder-GPTQ 

Octocoder GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Octocoder GPTQ (TheBloke/Octocoder-GPTQ)
๐ŸŒŸ Advertise your project ๐Ÿš€

Octocoder GPTQ Parameters and Internals

Model Type 
text-generation
Use Cases 
Areas:
Software development, Code generation
Applications:
Programming assistance, Coding instruction
Primary Use Cases:
Assisting programmers in writing code, Providing coding solutions based on instructions
Limitations:
May not provide optimal solutions for complex problems., Performance is dependent on the quality of the input prompt.
Considerations:
Preface input with 'Question: ' and finish it with 'Answer:'
Supported Languages 
programming_languages (80+ Programming languages)
Training Details 
Data Sources:
bigcode/commitpackft, bigcode/oasst-octopack
Data Volume:
1 trillion pretraining & 2M instruction tuning tokens
Methodology:
Instruction Tuning on CommitPackFT and OASST
Training Time:
Pretraining: 24 days, Instruction tuning: 4 hours
Hardware Used:
Pretraining: 512 Tesla A100 GPUs, Instruction tuning: 8 Tesla A100 GPUs
Model Architecture:
GPT-2 model with multi-query attention and Fill-in-the-Middle objective
Input Output 
Input Format:
Preface input with 'Question:' and finish with 'Answer:'
Accepted Modalities:
text
Output Format:
text generation
Performance Tips:
Use quality prompts to improve output relevance.
Release Notes 
Version:
initial
Notes:
Introduction of OctoCoder with instruction tuning based on StarCoder.
LLM NameOctocoder GPTQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/Octocoder-GPTQ 
Model CreatorBigCode
Base Model(s)  Octocoder   bigcode/octocoder
Model Size2.7b
Required VRAM9.2 GB
Updated2025-09-23
MaintainerTheBloke
Model Typegpt_bigcode
Model Files  9.2 GB
GPTQ QuantizationYes
Quantization Typegptq
Generates CodeYes
Model ArchitectureGPTBigCodeForCausalLM
Licensebigcode-openrail-m
Transformers Version4.31.0.dev0
Tokenizer ClassGPT2Tokenizer
Vocabulary Size49152
Torch Data Typefloat32
Activation Functiongelu

Best Alternatives to Octocoder GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Sqlcoder2 GPTQ0K / 9.2 GB1311
Sqlcoder GPTQ0K / 9.2 GB339
Starcoder Gptq0K / 9.2 GB11780
Redmond Hermes Coder GPTQ0K / 9.2 GB710
Starcoder Gptq Testing0K / 9.2 GB80
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Octocoder-GPTQ.

Rank the Octocoder GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51538 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124