NVIDIA Nemotron 3 Super 120B A12B FP8 by unsloth

 ยป  All LLMs  ยป  unsloth  ยป  NVIDIA Nemotron 3 Super 120B A12B FP8   URL Share it on

NVIDIA Nemotron 3 Super 120B A12B FP8 is an open-source language model by unsloth. Features: 120b LLM, VRAM: 128.4GB, Context: 256K, License: other.

  Arxiv:2512.20848   Arxiv:2512.20856 Base model:nvidia/nvidia-nemot... Base model:quantized:nvidia/nv...   Conversational   Custom code Dataset:nvidia/nemotron-post-t... Dataset:nvidia/nemotron-pre-tr...   De   En   Endpoints compatible   Es   Fr   It   Ja   Latent-moe   Modelopt   Mtp   Nemotron-3   Nemotron h   Nvidia   Pytorch   Region:us   Safetensors   Sharded   Tensorflow   Unsloth   Zh

NVIDIA Nemotron 3 Super 120B A12B FP8 Parameters and Internals

LLM NameNVIDIA Nemotron 3 Super 120B A12B FP8
Repository ๐Ÿค—https://huggingface.co/unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-FP8 
Base Model(s)  nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-BF16   nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-BF16
Model Size120b
Required VRAM128.4 GB
Updated2026-03-27
Maintainerunsloth
Model Typenemotron_h
Model Files  5.0 GB: 1-of-26   5.0 GB: 2-of-26   5.0 GB: 3-of-26   5.0 GB: 4-of-26   5.0 GB: 5-of-26   5.0 GB: 6-of-26   5.0 GB: 7-of-26   5.0 GB: 8-of-26   5.0 GB: 9-of-26   5.0 GB: 10-of-26   5.0 GB: 11-of-26   5.0 GB: 12-of-26   5.0 GB: 13-of-26   5.0 GB: 14-of-26   5.0 GB: 15-of-26   5.0 GB: 16-of-26   5.0 GB: 17-of-26   5.0 GB: 18-of-26   5.0 GB: 19-of-26   5.0 GB: 20-of-26   5.0 GB: 21-of-26   5.0 GB: 22-of-26   5.0 GB: 23-of-26   5.0 GB: 24-of-26   5.0 GB: 25-of-26   3.4 GB: 26-of-26
Supported Languagesen fr es it de ja zh
Model ArchitectureNemotronHForCausalLM
Licenseother
Context Length262144
Model Max Length262144
Transformers Version4.57.6
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<SPECIAL_999>
Vocabulary Size131072
Torch Data Typebfloat16

Best Alternatives to NVIDIA Nemotron 3 Super 120B A12B FP8

Best Alternatives
Context / RAM
Downloads
Likes
...motron 3 Super 120B A12B NVFP4256K / 80.3 GB1058108219
...Nemotron 3 Super 120B A12B FP8256K / 128.4 GB873745202
...uper 120B A12B BF16 Heretic V2256K / 241.4 GB2480

Rank the NVIDIA Nemotron 3 Super 120B A12B FP8 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51648 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260327b