Qwen3 4B Tcomanr Merge V2 by ertghiu256

 ยป  All LLMs  ยป  ertghiu256  ยป  Qwen3 4B Tcomanr Merge V2   URL Share it on

  Merged Model   Arxiv:2306.01708   Autotrain compatible Base model:ertghiu256/deepseek... Base model:ertghiu256/qwen-3-4... Base model:ertghiu256/qwen3-4b... Base model:ertghiu256/qwen3-he... Base model:ertghiu256/qwen3-ma... Base model:ertghiu256/qwen3-mu... Base model:huihui-ai/huihui-qw... Base model:polaris-project/pol... Base model:qwen/qwen3-4b-think... Base model:tesslate/uigen-t3-4... Base model:valiantlabs/qwen3-4... Base model:valiantlabs/qwen3-4...   Conversational   Endpoints compatible   Gguf   Q8   Quantized   Qwen3   Region:us   Safetensors   Sharded   Tensorflow

Qwen3 4B Tcomanr Merge V2 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Qwen3 4B Tcomanr Merge V2 (ertghiu256/Qwen3-4b-tcomanr-merge-v2)
๐ŸŒŸ Advertise your project ๐Ÿš€

Qwen3 4B Tcomanr Merge V2 Parameters and Internals

LLM NameQwen3 4B Tcomanr Merge V2
Repository ๐Ÿค—https://huggingface.co/ertghiu256/Qwen3-4b-tcomanr-merge-v2 
Base Model(s)  Polaris 4B Preview   Qwen 3 4B Mixture Of Thought   Qwen3 4B Thinking 2507   Qwen3 Multi Reasoner   Qwen3 Hermes 4B   ...epseek R1 0528 Distilled Qwen3   UIGEN T3 4B Preview MAX   Qwen3 Math Reasoner   Qwen3 4B Esper3   ...3 4B Thinking 2507 Abliterated   Qwen3 4B Code Reasoning   Qwen3 4B ShiningValiant3   POLARIS-Project/Polaris-4B-Preview   ertghiu256/qwen-3-4b-mixture-of-thought   Qwen/Qwen3-4B-Thinking-2507   ertghiu256/qwen3-multi-reasoner   ertghiu256/Qwen3-Hermes-4b   ertghiu256/deepseek-r1-0528-distilled-qwen3   Tesslate/UIGEN-T3-4B-Preview-MAX   ertghiu256/qwen3-math-reasoner   ValiantLabs/Qwen3-4B-Esper3   huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated   ertghiu256/qwen3-4b-code-reasoning   ValiantLabs/Qwen3-4B-ShiningValiant3
Merged ModelYes
Model Size4b
Required VRAM8 GB
Updated2025-09-23
Maintainerertghiu256
Model Typeqwen3
Model Files  8.1 GB   4.3 GB   5.0 GB: 1-of-2   3.0 GB: 2-of-2
GGUF QuantizationYes
Quantization Typeq8|gguf
Model ArchitectureQwen3ForCausalLM
Context Length262144
Model Max Length262144
Transformers Version4.51.3
Tokenizer ClassQwen2Tokenizer
Padding Token<|endoftext|>
Vocabulary Size151936
Torch Data Typefloat16
Errorsreplace

Best Alternatives to Qwen3 4B Tcomanr Merge V2

Best Alternatives
Context / RAM
Downloads
Likes
...wen3 4B Toolcalling Gguf Codex256K / 4.3 GB59413
...wen3 4B Thinking 2507 Hermes 3256K / 8.1 GB5411
Qwen3 4B Tcomanr Merge V2.2256K / 8 GB6712
...B Toolcall Gguf Llamacpp Codex256K / 4.3 GB822
Qwen3 4B 128K GGUF128K / 1.1 GB364924
Qwen3 4B GGUF40K / 1.1 GB3856778
ReasonableQwen3 4B40K / 2.5 GB71632
Qwen3 4B GGUF40K / 1.7 GB14166
Qwen3 Hermes 4B40K / 8.1 GB1422
...wen3 4B Thinking 2507 MLX 8bit256K / 4.3 GB6281507
Note: green Score (e.g. "73.2") means that the model is better than ertghiu256/Qwen3-4b-tcomanr-merge-v2.

Rank the Qwen3 4B Tcomanr Merge V2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51538 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124