| Model Type | | instruct fine-tuned model |
|
| Use Cases |
| Areas: | | local intelligence, on-device computing, at-the-edge use cases |
|
| Applications: | |
| Primary Use Cases: | | non-commercial research purposes |
|
| Limitations: | | Only for non-commercial research purposes |
|
|
| Additional Notes | | Trained with a 128k context window utilizing interleaved sliding-window attention. |
|
| Supported Languages | | en (English), fr (French), de (German), es (Spanish), it (Italian), pt (Portuguese), zh (Chinese), ja (Japanese), ru (Russian), ko (Korean) |
|
| Training Details |
| Context Length: | |
| Model Architecture: | |
|
| Input Output |
| Input Format: | | V3-Tekken tokenizer format |
|
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use Mistral Inference or vLLM for optimized performance. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Release of Ministral-8B-Instruct-2410 under Mistral Research License |
|
|
|