Swallow 70B GPTQ is an open-source language model by TheBloke. Features: 70b LLM, VRAM: 35.7GB, Context: 4K, License: llama2, Quantized, LLM Explorer Score: 0.11.
The Swallow 70B model has been enhanced for Japanese with significant improvements in JCommonsenseQA, JEMHopQA, NIILC, JSQuAD, MGSM, and WMT20 benchmark tasks.
Supported Languages
Japanese (high), English (high)
Training Details
Data Sources:
Japanese Wikipedia, RefinedWeb, Swallow Corpus, The Pile
Methodology:
The Swallow model has undergone continuous pre-training from the Llama 2 family, primarily with the addition of Japanese language data.
Model Architecture:
Refer to LLaMA-2 technical report for details on the model architecture.
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Swallow-70B-GPTQ.
Rank the Swallow 70B GPTQ Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.