Swallow 70B AWQ is an open-source language model by TheBloke. Features: 70b LLM, VRAM: 37GB, Context: 4K, License: llama2, Quantized, LLM Explorer Score: 0.11.
Swallow model versions range from 7B to 70B, focusing mainly on Japanese-English bilingual tasks. The model employs advanced quantization techniques for efficient inference.
Supported Languages
en (English), ja (Japanese)
Training Details
Data Sources:
Japanese Wikipedia, RefinedWeb, Swallow Corpus, The Pile
Methodology:
Continual pre-training and instruction tuning using datasets like Anthropic HH-RLHF and Databricks Dolly 15-k.
Context Length:
4096
Model Architecture:
Llama-2 architecture based, with a Japanese vocabulary extended tokenizer.
Input Output
Input Format:
{prompt}
Accepted Modalities:
text
Output Format:
Generated text
Performance Tips:
Efficient text generation with Japanese input due to optimized tokenizer. For fast inference, GPU usage is recommended.
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Swallow-70B-AWQ.
Rank the Swallow 70B AWQ Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.