Swallow 70B GGUF by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Swallow 70B GGUF   URL Share it on

Swallow 70B GGUF is an open-source language model by TheBloke. Features: 70b LLM, VRAM: 29.4GB, License: llama2, Quantized, LLM Explorer Score: 0.11.

Base model:quantized:tokyotech... Base model:tokyotech-llm/swall...   En   Gguf   Ja   Llama   Quantized   Region:us

Swallow 70B GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Swallow 70B GGUF (TheBloke/Swallow-70B-GGUF)
๐ŸŒŸ Advertise your project ๐Ÿš€

Swallow 70B GGUF Parameters and Internals

Model Type 
llama
Use Cases 
Limitations:
Models are still in the early stages of research and development and have not been tuned for alignment with human intent and safety considerations.
Additional Notes 
The model features a tokenizer with a broadened vocabulary for Japanese data, enhancing text representation efficiency.
Supported Languages 
Japanese (Proficient), English (Proficient)
Training Details 
Data Sources:
Japanese Wikipedia, RefinedWeb, Swallow Corpus, The Pile
Methodology:
Continual pre-training from Llama 2 with addition of Japanese language data
Model Architecture:
LLaMA-2 architecture
LLM NameSwallow 70B GGUF
Repository ๐Ÿค—https://huggingface.co/TheBloke/Swallow-70B-GGUF 
Model NameSwallow 70B
Model Creatortokyotech-llm
Base Model(s)  Swallow 70B Hf   tokyotech-llm/Swallow-70b-hf
Model Size70b
Required VRAM29.4 GB
Updated2026-03-29
MaintainerTheBloke
Model Typellama
Model Files  29.4 GB   36.3 GB   33.3 GB   30.0 GB   39.0 GB   41.5 GB   39.2 GB   47.6 GB   48.9 GB   47.6 GB
Supported Languagesen ja
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureAutoModel
Licensellama2

Best Alternatives to Swallow 70B GGUF

Best Alternatives
Context / RAM
Downloads
Likes
...us Qwen3 R1 Llama Distill GGUF0K / 0.8 GB2112
KafkaLM 70B German V0.1 GGUF0K / 25.5 GB240256
...gekit Passthrough Yqhuxcv GGUF0K / 16.9 GB1100
CodeLlama 70B Instruct GGUF0K / 25.5 GB215960
CodeLlama 70B Python GGUF0K / 25.5 GB159644
Meta Llama 3 70B Instruct GGUF0K / 26.4 GB1334
DAD Model V2 70B Q40K / 42.5 GB80
CodeLlama 70B Hf GGUF0K / 25.5 GB55842
Swallow 70B Instruct GGUF0K / 29.4 GB15509
Llama 2 70B Guanaco QLoRA GGUF0K / 29.3 GB190
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Swallow-70B-GGUF.

Rank the Swallow 70B GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a