Mistralai Mistral Nemo Base 2407 is an open-source language model by SillyTilly. Features: 12.2b LLM, VRAM: 24.5GB, License: apache-2.0, LLM Explorer Score: 0.15.
Mistralai Mistral Nemo Base 2407 Parameters and Internals
Model Type
generative, transformer
Additional Notes
Mistral-Nemo-Base-2407 is a pretrained base model without moderation mechanisms.
Supported Languages
en (Proficient), fr (Proficient), de (Proficient), es (Proficient), it (Proficient), pt (Proficient), ru (Proficient), zh (Proficient), ja (Proficient)
Training Details
Methodology:
Pre-trained generative text model using a 128k context window and trained on a large proportion of multilingual and code data.
Context Length:
128000
Model Architecture:
Transformer model with 40 layers, dimension 5,120, head dimension 128, hidden dimension 14,436, SwiGLU activation function, 32 heads, 8 kv-heads, vocabulary size 128k with rotary embeddings (theta = 1M).
Note: green Score (e.g. "73.2") means that the model is better than SillyTilly/mistralai_Mistral-Nemo-Base-2407.
Rank the Mistralai Mistral Nemo Base 2407 Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52758 in total.