Fine Tuned SmolLM2 360M With On Combined Instruction Dataset by akhilfau

 ยป  All LLMs  ยป  akhilfau  ยป  Fine Tuned SmolLM2 360M With On Combined Instruction Dataset   URL Share it on

Fine Tuned SmolLM2 360M With On Combined Instruction Dataset is an open-source language model by akhilfau. Features: 360m LLM, VRAM: 1.4GB, Context: 8K, License: apache-2.0, Instruction-Based, LLM Explorer Score: 0.19.

  Autotrain compatible Base model:finetune:huggingfac... Base model:huggingfacetb/smoll...   Conversational   Dataset:camel-ai/physics   Education   En   Endpoints compatible   Instruct   Llama   Low-resource-llm   Physics,   Region:us   Safetensors   Small-language-model,   Smollm,

Fine Tuned SmolLM2 360M With On Combined Instruction Dataset Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Fine Tuned SmolLM2 360M With On Combined Instruction Dataset (akhilfau/fine-tuned-smolLM2-360M-with-on-combined_Instruction_dataset)
๐ŸŒŸ Advertise your project ๐Ÿš€

Fine Tuned SmolLM2 360M With On Combined Instruction Dataset Parameters and Internals

LLM NameFine Tuned SmolLM2 360M With On Combined Instruction Dataset
Repository ๐Ÿค—https://huggingface.co/akhilfau/fine-tuned-smolLM2-360M-with-on-combined_Instruction_dataset 
Base Model(s)  SmolLM2 360M   HuggingFaceTB/SmolLM2-360M
Model Size360m
Required VRAM1.4 GB
Updated2025-09-29
Maintainerakhilfau
Model Typellama
Instruction-BasedYes
Model Files  1.4 GB
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length8192
Model Max Length8192
Transformers Version4.49.0
Tokenizer ClassGPT2Tokenizer
Padding Token<|im_end|>
Vocabulary Size49152
Torch Data Typefloat32

Best Alternatives to Fine Tuned SmolLM2 360M With On Combined Instruction Dataset

Best Alternatives
Context / RAM
Downloads
Likes
SmolLM2 360M Instruct8K / 0.7 GB446881185
Smollm3 720prms8K / 0.7 GB460
ProseFlow V1 360M Instruct8K / 0.7 GB80
SmolLM2 Rethink 360M8K / 1.4 GB71
SolaraV2 Coder 05118K / 0.7 GB51
BrainrotLM Assistant 362M8K / 0 GB70
SmolLM2 CoT 360M8K / 1.4 GB819
... Cpt Fineweb Norwegian Nynorsk8K / 1.4 GB60
SmolLM2 360M Instruct Ita8K / 0.7 GB16050
SmolLM2 360M Instruct8K / 0.7 GB16414
Note: green Score (e.g. "73.2") means that the model is better than akhilfau/fine-tuned-smolLM2-360M-with-on-combined_Instruction_dataset.

Rank the Fine Tuned SmolLM2 360M With On Combined Instruction Dataset Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a