NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 by trohrbaugh

 ยป  All LLMs  ยป  trohrbaugh  ยป  NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2   URL Share it on

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 is an open-source language model by trohrbaugh. Features: 120b LLM, VRAM: 241.4GB, Context: 256K, License: other, LLM Explorer Score: 0.3.

  Arxiv:2512.20848   Arxiv:2512.20856   Abliterated   Ara   Conversational   Custom code Dataset:nvidia/nemotron-post-t... Dataset:nvidia/nemotron-pre-tr...   De   Decensored   En   Endpoints compatible   Es   Fr   Heretic   It   Ja   Latent-moe   Mtp   Nemotron-3   Nemotron h   Nvidia   Pytorch   Region:us   Safetensors   Sharded   Tensorflow   Uncensored   Zh

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 (trohrbaugh/NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-v2)
๐ŸŒŸ Advertise your project ๐Ÿš€

NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Parameters and Internals

LLM NameNVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2
Repository ๐Ÿค—https://huggingface.co/trohrbaugh/NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-v2 
Model Size120b
Required VRAM241.4 GB
Updated2026-04-03
Maintainertrohrbaugh
Model Typenemotron_h
Model Files  10.0 GB: 1-of-25   10.0 GB: 2-of-25   10.0 GB: 3-of-25   10.0 GB: 4-of-25   10.0 GB: 5-of-25   10.0 GB: 6-of-25   10.0 GB: 7-of-25   10.0 GB: 8-of-25   10.0 GB: 9-of-25   10.0 GB: 10-of-25   9.9 GB: 11-of-25   10.0 GB: 12-of-25   10.0 GB: 13-of-25   9.9 GB: 14-of-25   10.0 GB: 15-of-25   10.0 GB: 16-of-25   9.9 GB: 17-of-25   10.0 GB: 18-of-25   10.0 GB: 19-of-25   9.9 GB: 20-of-25   10.0 GB: 21-of-25   10.0 GB: 22-of-25   9.9 GB: 23-of-25   10.0 GB: 24-of-25   1.9 GB: 25-of-25
Supported Languagesen fr es it de ja zh
Model ArchitectureNemotronHForCausalLM
Licenseother
Context Length262144
Model Max Length262144
Transformers Version5.3.0
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|im_end|>
Vocabulary Size131072

Best Alternatives to NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2

Best Alternatives
Context / RAM
Downloads
Likes
...on 3 Super 120B A12B Base BF161024K / 209.5 GB1129026
...motron 3 Super 120B A12B NVFP4256K / 80.3 GB1283305234
...emotron 3 Super 120B A12B BF16256K / 194.6 GB163028310
...Nemotron 3 Super 120B A12B FP8256K / 128.4 GB873745202
...Nemotron 3 Super 120B A12B FP8256K / 128.4 GB51379
...DIA Nemotron 3 Super 120B A12B256K / 214.5 GB4512
...motron 3 Super 120B A12B NVFP4256K / 80.3 GB4608222
...20B A12B BF16 REAP 50pct Draft256K / 128.5 GB816
Note: green Score (e.g. "73.2") means that the model is better than trohrbaugh/NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-v2.

Rank the NVIDIA Nemotron 3 Super 120B A12B BF16 Heretic V2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 52721 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Check out Ag3ntum โ€” our secure, self-hosted AI agent for server management.
Release v20260328a