MistralLite by AWS

 ยป  All LLMs  ยป  AWS  ยป  MistralLite   URL Share it on

MistralLite is an open-source language model by AWS. Features: LLM, VRAM: 14.4GB, Context: 32K, License: apache-2.0, HF Score: 51.5, LLM Explorer Score: 0.19, Arc: 59.6, HellaSwag: 81.8, MMLU: 50.9, TruthfulQA: 37.9, WinoGrande: 77.4, GSM8K: 1.1.

  Mistral   Pytorch   Region:us   Sharded
Model Card on HF ๐Ÿค—: https://huggingface.co/amazon/MistralLite 

MistralLite Benchmarks

MistralLite (amazon/MistralLite)
๐ŸŒŸ Advertise your project ๐Ÿš€

MistralLite Parameters and Internals

Model Type 
Language Model, Text Generation
Use Cases 
Areas:
Research, Commercial applications
Applications:
Long context retrieval, Summarization, Question-answering
Primary Use Cases:
Long context line and topic retrieval, Summarization, Question-answering
Limitations:
Performance may vary based on specific long context tasks and input lengths.
Considerations:
Use prompt templates for effective outcomes.
Additional Notes 
MistralLite supports various deployment methods suitable for different environments. It requires initial setup but offers improved performance for long context tasks.
Supported Languages 
English (Proficient)
Training Details 
Data Sources:
SLidingEncoder and Decoder (SLED), (Long) Natural Questions (NQ), OpenAssistant Conversations Dataset (OASST1)
Methodology:
Utilized an adapted Rotary Embedding and sliding window during fine-tuning
Context Length:
32000
Model Architecture:
Fine-tuned version of the Mistral-7B-v0.1 model using adaptations for long context handling.
Input Output 
Input Format:
Prompt templates such as '<|prompter|>What are the main challenges to support a long context for LLM?~~<|assistant|>'
Accepted Modalities:
text
Output Format:
Generated text responses aligned with input prompts
Performance Tips:
Use prompt templates for optimal model performance.
LLM NameMistralLite
Repository ๐Ÿค—https://huggingface.co/amazon/MistralLite 
Required VRAM14.4 GB
Updated2026-04-11
MaintainerAWS
Model Typemistral
Model Files  9.9 GB: 1-of-2   4.5 GB: 2-of-2
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.34.0
Tokenizer ClassLlamaTokenizer
Padding Token[PAD]
Vocabulary Size32003
Torch Data Typebfloat16

Quantized Models of the MistralLite

Model
Likes
Downloads
VRAM
MistralLite 7B GGUF415973 GB
MistralLite 7B GGUF12182 GB
MistralLite 7B AWQ894 GB
MistralLite 7B GPTQ3224 GB

Best Alternatives to MistralLite

Best Alternatives
Context / RAM
Downloads
Likes
Krutrim 2 Instruct1000K / 49.3 GB14736
Ft V1 Violet1000K / 24.5 GB50
Mistral Large Instruct 2407128K / 226.7 GB7491859
Tiny Random MistralForCausalLM128K / 0 GB32521
Winterreise M732K / 14.4 GB00
Frostwind V2.1 M732K / 14.4 GB00
K2S3 V0.132K / 28.7 GB60
MistralLite32K / 14.4 GB61777430
...ydaz Web AI Reasoner BaseModel32K / 14.4 GB01
Tess XS V1.3 Yarn 128K32K / 14.5 GB332013
Note: green Score (e.g. "73.2") means that the model is better than amazon/MistralLite.

Rank the MistralLite Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a