NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 by trohrbaugh

 »  All LLMs  »  trohrbaugh  »  NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2   URL Share it on

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 is an open-source language model by trohrbaugh. Features: 120b LLM, VRAM: 241.4GB, Context: 256K, License: other, LLM Explorer Score: 0.3.

  Arxiv:2512.20848   Arxiv:2512.20856   Abliterated   Ara   Conversational   Custom code Dataset:nvidia/nemotron-post-t... Dataset:nvidia/nemotron-pre-tr...   De   Decensored   En   Endpoints compatible   Es   Fr   Heretic   It   Ja   Latent-moe   Mtp   Nemotron-3   Nemotron h   Nvidia   Pytorch   Region:us   Safetensors   Sharded   Tensorflow   Uncensored   Zh

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Parameters and Internals

LLM NameNVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2
Repository 🤗https://huggingface.co/trohrbaugh/NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-v2 
Model Size120b
Required VRAM241.4 GB
Updated2026-05-11
Maintainertrohrbaugh
Model Typenemotron_h
Model Files  10.0 GB: 1-of-25   10.0 GB: 2-of-25   10.0 GB: 3-of-25   10.0 GB: 4-of-25   10.0 GB: 5-of-25   10.0 GB: 6-of-25   10.0 GB: 7-of-25   10.0 GB: 8-of-25   10.0 GB: 9-of-25   10.0 GB: 10-of-25   9.9 GB: 11-of-25   10.0 GB: 12-of-25   10.0 GB: 13-of-25   9.9 GB: 14-of-25   10.0 GB: 15-of-25   10.0 GB: 16-of-25   9.9 GB: 17-of-25   10.0 GB: 18-of-25   10.0 GB: 19-of-25   9.9 GB: 20-of-25   10.0 GB: 21-of-25   10.0 GB: 22-of-25   9.9 GB: 23-of-25   10.0 GB: 24-of-25   1.9 GB: 25-of-25
Supported Languagesen fr es it de ja zh
Model ArchitectureNemotronHForCausalLM
Licenseother
Context Length262144
Model Max Length262144
Transformers Version5.3.0
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|im_end|>
Vocabulary Size131072

Best Alternatives to NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2

Best Alternatives
Context / RAM
Downloads
Likes
...on 3 Super 120B A12B Base BF161024K / 209.5 GB2267830
...emotron 3 Super 120B A12B BF16256K / 194.6 GB729273344
...motron 3 Super 120B A12B NVFP4256K / 80.3 GB894238290
...Nemotron 3 Super 120B A12B FP8256K / 128.4 GB370440244
...Nemotron 3 Super 120B A12B FP8256K / 128.4 GB14869
...DIA Nemotron 3 Super 120B A12B256K / 214.5 GB17233
...motron 3 Super 120B A12B NVFP4256K / 80.3 GB4608222
... Super 64B A12B Math REAP BF16256K / 128.6 GB6571
...20B A12B BF16 REAP 50pct Draft256K / 128.5 GB2046
...emotron 3 Super 120B A12B 5bit256K / 83.1 GB28802
Note: green Score (e.g. "73.2") means that the model is better than trohrbaugh/NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-v2.

Rank the NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 53640 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum — our secure, self-hosted AI agent for server management.
Release v20260328a