Mixture Of Recursions 360M by sudeshmu

 ยป  All LLMs  ยป  sudeshmu  ยป  Mixture Of Recursions 360M   URL Share it on

  Arxiv:2507.10524   Adaptive-computation   Autotrain compatible Base model:finetune:microsoft/... Base model:microsoft/dialogpt-...   Custom code Dataset:huggingfacetb/smollm-c...   Early-exiting   Efficient-inference   En   Language-model   Llama   Mixture-of-recursions   Mor llama   Pytorch   Region:us

Mixture Of Recursions 360M Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mixture Of Recursions 360M (sudeshmu/mixture-of-recursions-360m)
๐ŸŒŸ Advertise your project ๐Ÿš€

Mixture Of Recursions 360M Parameters and Internals

LLM NameMixture Of Recursions 360M
Repository ๐Ÿค—https://huggingface.co/sudeshmu/mixture-of-recursions-360m 
Base Model(s)  DialoGPT Medium   microsoft/DialoGPT-medium
Model Size360m
Required VRAM0.5 GB
Updated2025-07-29
Maintainersudeshmu
Model Typemor_llama
Model Files  0.5 GB
Supported Languagesen
Model ArchitectureMoRLlamaForCausalLM
Licensemit
Context Length1024
Model Max Length1024
Transformers Version4.50.0
Tokenizer ClassLlamaTokenizer
Vocabulary Size49152
Torch Data Typebfloat16

Rank the Mixture Of Recursions 360M Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50200 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124