| Model Type | |
| Use Cases |
| Areas: | | Synthetic Data Generation, building and customizing LLMs |
|
| Applications: | | Chat applications, AI assistant |
|
| Primary Use Cases: | |
| Limitations: | | Amplifies biases from training data, may generate socially undesirable text |
|
|
| Supported Languages | | languages_supported (Multilingual), proficiency_levels () |
|
| Training Details |
| Data Sources: | | 9 trillion tokens of English based texts, 50+ natural languages, and 40+ coding languages |
|
| Methodology: | | Supervised Fine-tuning (SFT), Direct Preference Optimization (DPO), Reward-aware Preference Optimization (RPO), Grouped-Query Attention (GQA), Rotary Position Embeddings (RoPE) |
|
| Context Length: | |
| Training Time: | |
| Model Architecture: | |
|
| Safety Evaluation |
| Methodologies: | | Adversarial testing via Garak, AEGIS content safety evaluation, Human Content Red Teaming |
|
| Risk Categories: | | Toxic language, unsafe content, societal biases |
|
| Ethical Considerations: | | NVIDIA believes Trustworthy AI is a shared responsibility. |
|
|
| Input Output |
| Input Format: | | Single Turn: System User {prompt} Assistant; Multi-Turn: User {prompt 1} Assistant {response 1} User {prompt 2} Assistant {response 2}... |
|
| Output Format: | |
|