Merged16 Mistral 32 Large Grpo 5 Weak3 by Nitish035

 ยป  All LLMs  ยป  Nitish035  ยป  Merged16 Mistral 32 Large Grpo 5 Weak3   URL Share it on

  4bit   Autotrain compatible Base model:finetune:unsloth/mi... Base model:unsloth/mistral-7b-...   Conversational   En   Endpoints compatible   Instruct   Mistral   Quantized   Region:us   Safetensors   Sharded   Tensorflow   Unsloth

Merged16 Mistral 32 Large Grpo 5 Weak3 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Merged16 Mistral 32 Large Grpo 5 Weak3 (Nitish035/merged16_mistral_32_large_grpo-5-weak3)
๐ŸŒŸ Advertise your project ๐Ÿš€

Merged16 Mistral 32 Large Grpo 5 Weak3 Parameters and Internals

LLM NameMerged16 Mistral 32 Large Grpo 5 Weak3
Repository ๐Ÿค—https://huggingface.co/Nitish035/merged16_mistral_32_large_grpo-5-weak3 
Base Model(s)  unsloth/mistral-7b-instruct-v0.3-bnb-4bit   unsloth/mistral-7b-instruct-v0.3-bnb-4bit
Model Size7b
Required VRAM14.5 GB
Updated2025-07-29
MaintainerNitish035
Model Typemistral
Instruction-BasedYes
Model Files  5.0 GB: 1-of-3   5.0 GB: 2-of-3   4.5 GB: 3-of-3
Supported Languagesen
Quantization Type4bit
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.54.0
Tokenizer ClassLlamaTokenizer
Padding Token[control_768]
Vocabulary Size32768
Torch Data Typefloat16

Best Alternatives to Merged16 Mistral 32 Large Grpo 5 Weak3

Best Alternatives
Context / RAM
Downloads
Likes
...al Nemo Instruct 2407 Bnb 4bit128K / 8.3 GB58523929
...tral 7B Instruct V0.3 Bnb 4bit32K / 4.1 GB5269428
Mistral 7B Instruct V0.2 Fp1632K / 14.4 GB260
...tral 7B Instruct V0.2 Bnb 4bit32K / 4.1 GB2250133
Mistral 7B Instruct V0.2 4bit32K / 4.3 GB1641
Mixtral V0.3 Full 16bit32K / 14.5 GB50
... 7B Instruct V0.3ContinuedFine32K / 14.5 GB160
Mistral 7B Instruct V0.2 8bit32K / 7.6 GB71
...eMistral 7B Instruct Ead 16bit32K / 14.5 GB380
NaturalLM 7B Instruct32K / 14.5 GB310
Note: green Score (e.g. "73.2") means that the model is better than Nitish035/merged16_mistral_32_large_grpo-5-weak3.

Rank the Merged16 Mistral 32 Large Grpo 5 Weak3 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50200 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124