Model trained on unfiltered internet data, may contain objectionable content.
Supported Languages
en (high), zh (high), ja (high), de (high)
Training Details
Data Sources:
synthetic dataset generated using large context windows, retrieval-augmented generation, and knowledge graph integration
Data Volume:
20 billion tokens
Methodology:
fine-tuning using a synthesis dataset
Context Length:
1000000
Training Time:
< 1 day on 16 nodes of 8*A100-80G
Hardware Used:
16 nodes of 8*A100-80G GPUs
Input Output
Input Format:
Accepts text and image modalities.
Accepted Modalities:
text, image
Performance Tips:
Use a standardized implementation for inference to avoid performance degradation. For fewer hallucinations, use top_p=0.8 and temperature=0.3, or temperature=0.2.
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.