Mamba 2.8B Chat No Robots by clibrain

 ยป  All LLMs  ยป  clibrain  ยป  Mamba 2.8B Chat No Robots   URL Share it on

  Arxiv:2203.02155   Conversational Dataset:huggingfaceh4/no robot...   En   Endpoints compatible   Pytorch   Region:us   Tensorboard

Mamba 2.8B Chat No Robots Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mamba 2.8B Chat No Robots (clibrain/mamba-2.8b-chat-no_robots)
๐ŸŒŸ Advertise your project ๐Ÿš€

Mamba 2.8B Chat No Robots Parameters and Internals

Model Type 
text generation
Additional Notes 
The model is inspired by Mamba state space model architecture with focus on efficient handling of information-dense data.
Supported Languages 
English (en)
Training Details 
Data Sources:
HuggingFaceH4/no_robots
Methodology:
fine-tuning on instruction-following data
Model Architecture:
Structured state space model with hardware-aware design
Input Output 
Input Format:
text-based prompts with chat template
Accepted Modalities:
text
Output Format:
generated text responses
LLM NameMamba 2.8B Chat No Robots
Repository ๐Ÿค—https://huggingface.co/clibrain/mamba-2.8b-chat-no_robots 
Model Size2.8b
Required VRAM5.5 GB
Updated2025-08-21
Maintainerclibrain
Model Files  5.5 GB
Supported Languagesen
Model ArchitectureAutoModel
Licensewtfpl
Tokenizer ClassGPTNeoXTokenizer
Padding Token<|endoftext|>
Vocabulary Size50277

Best Alternatives to Mamba 2.8B Chat No Robots

Best Alternatives
Context / RAM
Downloads
Likes
Mamba 2.8B0K / 11.1 GB11936148
Mamba 2.8B Slimpj0K / 11.1 GB324128
Synatra Mamba Ko 2.8B0K / 5.8 GB11
Mamba Chat 2.8B0K / 5.5 GB13
Bagel DPO 2.8B V0.20K / 11.1 GB120
Mamba 2.8B Instruct Openhermes0K / 5.5 GB471
Mamba 2.8B CyberSec0K / 5.5 GB29
Bagel 2.8B V0.20K / 11.1 GB17
...a Financial Headline Sentiment0K / 5.5 GB21
Ct2fast Pythia 2.8B0K / 5.5 GB31
Note: green Score (e.g. "73.2") means that the model is better than clibrain/mamba-2.8b-chat-no_robots.

Rank the Mamba 2.8B Chat No Robots Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50804 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124