Fireball MathMistral Nemo Base 2407 is an open-source language model by EpistemeAI. Features: 12.2b LLM, VRAM: 24.5GB, Context: 1000K, License: apache-2.0, Quantized, LLM Explorer Score: 0.15.
Fireball MathMistral Nemo Base 2407 Parameters and Internals
Model Type
text generation
Use Cases
Areas:
Research, Education, Mathematics
Applications:
Math problem solving
Primary Use Cases:
Provide better math responses than previous models
Additional Notes
Mistral-Nemo-Base-2407 is a pretrained base model and does not have any moderation mechanisms.
Supported Languages
en (Advanced)
Training Details
Data Sources:
meta-math/MetaMathQA
Methodology:
Supervised fine-tuning
Context Length:
128000
Model Architecture:
Transformer with 40 layers, 5,120 dim, 14,436 hidden dim, SwiGLU activation function. Number of heads: 32, number of kv-heads: 8 (GQA), Vocabulary size: 128k with Rotary embeddings (theta = 1M).
Note: green Score (e.g. "73.2") means that the model is better than EpistemeAI/Fireball-MathMistral-Nemo-Base-2407.
Rank the Fireball MathMistral Nemo Base 2407 Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.