Galactica 30B Evol Instruct 70K GPTQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Galactica 30B Evol Instruct 70K GPTQ   URL Share it on

  Arxiv:2304.12244   4-bit   Alpaca   Autotrain compatible Dataset:wizardlm/wizardlm evol...   Galactica   Gptq   Instruct   Opt   Quantized   Region:us   Safetensors   Wizardlm

Galactica 30B Evol Instruct 70K GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Galactica 30B Evol Instruct 70K GPTQ (TheBloke/galactica-30B-evol-instruct-70K-GPTQ)
๐ŸŒŸ Advertise your project ๐Ÿš€

Galactica 30B Evol Instruct 70K GPTQ Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
Research
Applications:
Scientific computing
Primary Use Cases:
Research in language models applied to scientific domains
Limitations:
Not suitable for production use without safeguards due to potential inaccuracies
Considerations:
Ensure proper verification of generated information for scientific accuracy.
Additional Notes 
Quantised to 4bit for improved efficiency.
Training Details 
Data Sources:
WizardLM/WizardLM_evol_instruct_70k
Data Volume:
106 billion tokens
Methodology:
Fine-tuned using the Evol-Instruct framework.
Context Length:
1024
Training Time:
196 hours using 16 A100 80GB GPUs
Hardware Used:
16 A100 80GB GPUs
Model Architecture:
GPT-Q (quantized to 4bit)
Input Output 
Input Format:
### Instruction: prompt ### Response:
Accepted Modalities:
text
Output Format:
Text
Performance Tips:
Use device='cuda:0' for better performance.
LLM NameGalactica 30B Evol Instruct 70K GPTQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/galactica-30B-evol-instruct-70K-GPTQ 
Base Model(s)  ...alactica 30B Evol Instruct 70K   GeorgiaTechResearchInstitute/galactica-30b-evol-instruct-70k
Model Size30b
Required VRAM16.3 GB
Updated2025-09-23
MaintainerTheBloke
Model Typeopt
Instruction-BasedYes
Model Files  16.3 GB
GPTQ QuantizationYes
Quantization Typegptq
Model ArchitectureOPTForCausalLM
Licenseother
Context Length2048
Model Max Length2048
Transformers Version4.29.2
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size50000
Torch Data Typebfloat16
Activation Functiongelu

Best Alternatives to Galactica 30B Evol Instruct 70K GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
...alactica 30B Evol Instruct 70K2K / 60.1 GB161823
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/galactica-30B-evol-instruct-70K-GPTQ.

Rank the Galactica 30B Evol Instruct 70K GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51538 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124