Swallow 70B GGUF by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Swallow 70B GGUF   URL Share it on

Base model:quantized:tokyotech... Base model:tokyotech-llm/swall...   En   Gguf   Ja   Llama   Quantized   Region:us

Swallow 70B GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
๐ŸŒŸ Advertise your project ๐Ÿš€

Swallow 70B GGUF Parameters and Internals

Model Type 
llama
Use Cases 
Limitations:
Models are still in the early stages of research and development and have not been tuned for alignment with human intent and safety considerations.
Additional Notes 
The model features a tokenizer with a broadened vocabulary for Japanese data, enhancing text representation efficiency.
Supported Languages 
Japanese (Proficient), English (Proficient)
Training Details 
Data Sources:
Japanese Wikipedia, RefinedWeb, Swallow Corpus, The Pile
Methodology:
Continual pre-training from Llama 2 with addition of Japanese language data
Model Architecture:
LLaMA-2 architecture
LLM NameSwallow 70B GGUF
Repository ๐Ÿค—https://huggingface.co/TheBloke/Swallow-70B-GGUF 
Model NameSwallow 70B
Model Creatortokyotech-llm
Base Model(s)  Swallow 70B Hf   tokyotech-llm/Swallow-70b-hf
Model Size70b
Required VRAM29.4 GB
Updated2025-06-09
MaintainerTheBloke
Model Typellama
Model Files  29.4 GB   36.3 GB   33.3 GB   30.0 GB   39.0 GB   41.5 GB   39.2 GB   47.6 GB   48.9 GB   47.6 GB
Supported Languagesen ja
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureAutoModel
Licensellama2
Swallow 70B GGUF (TheBloke/Swallow-70B-GGUF)

Best Alternatives to Swallow 70B GGUF

Best Alternatives
Context / RAM
Downloads
Likes
CodeLlama 70B Instruct GGUF0K / 25.5 GB269957
...gekit Passthrough Yqhuxcv GGUF0K / 16.9 GB110
Meta Llama 3 70B Instruct GGUF0K / 26.4 GB4913
CodeLlama 70B Python GGUF0K / 25.5 GB107643
KafkaLM 70B German V0.1 GGUF0K / 25.5 GB116738
DAD Model V2 70B Q40K / 42.5 GB110
CodeLlama 70B Hf GGUF0K / 25.5 GB52742
Llama 2 70B Guanaco QLoRA GGUF0K / 29.3 GB140
Meditron 70B GGUF0K / 29.3 GB55420
Swallow 70B Instruct GGUF0K / 29.4 GB3208
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Swallow-70B-GGUF.

Rank the Swallow 70B GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 48046 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124