| Model Type | | Mixture of Experts (MoE), Bilingual |
|
| Use Cases |
| Areas: | | Research, Commercial Applications |
|
| Limitations: | |
| Considerations: | | Engagement with the community for guardrails |
|
|
| Additional Notes | | The AquilaMoE Instruct model is a demonstration of easy fine-tuning for compelling performance |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | RedPajama-Data-V2, falcon-refinedweb, C4, Pile, WuDaoCorporaText, ChineseWebText |
|
| Data Volume: | |
| Methodology: | | EfficientScale with Scale-Up and Scale-Out strategies |
|
| Context Length: | |
|