| Model Type | | Mixture-Of-Experts, language model, text generation |
|
| Use Cases |
| Areas: | | coding, medical, bio, cybersecurity |
|
| Applications: | | programming support, medical inference, long-context data handling, transformer model tasks |
|
| Primary Use Cases: | | chat models, roleplay, instructional tasks, function calling |
|
|
| Additional Notes | | The MEGA_MIND 24b CyberSeries integrates diversified expert models into a unified framework. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | OpenOrca dataset, Mistral-7B-v0.2, PubMed Central Open Access |
|
| Data Volume: | |
| Methodology: | | Mixture of Experts, Reinforcement Learning from AI Feedback |
|
| Context Length: | |
| Model Architecture: | | Mixture of Experts, Transformer-based |
|
|