| Model Type | | bilingual, text generation, instruction fine-tuned, decoder, causal-lm |
|
| Use Cases |
| Areas: | |
| Applications: | | Arabic NLP, Chat applications, Sentiment analysis, Summarization |
|
| Primary Use Cases: | | Chat assistants, Arabic cultural phenomena, Quantitative studies, Mechanistic interpretability |
|
| Limitations: | | Malicious use, Sensitive Information, Generalization across all languages, High-stakes decisions |
|
| Considerations: | | Models should comply with applicable laws and not violate sensitive information handling. |
|
|
| Additional Notes | | Extensible approaches for adapting low and medium resource languages. |
|
| Supported Languages | | Arabic (MSA) (bilingual), English (bilingual) |
|
| Training Details |
| Data Sources: | | Web, Code, Books, Scientific, Synthetic |
|
| Data Volume: | |
| Methodology: | | Auto-regressive, Transformer-based, Decoder-only |
|
| Context Length: | |
| Hardware Used: | | Cerebras CS-2 Wafer-Scale Engines, Condor Galaxy supercomputer |
|
| Model Architecture: | | SwiGLU, ALiBi position encoding, RoPE, Grouped Query Attention |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Enable `trust_remote_code=True` when loading the model. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Initial release of Jais Family models ranging from 590M to 70B parameters. |
|
|
|