| Model Type | | large, decoder-only, transformer, autoregressive | 
 | 
| Use Cases | 
| Areas: | | research, evaluation of Large Language Models capabilities | 
 |  | Primary Use Cases: | | Validating the model and collecting feedback on Large Language Models. | 
 |  | Limitations: | | Bias, Safety, Generation diversity, Hallucination, Overrepresentation of some viewpoints, Discriminatory language, Inaccurate information generation, Repetitive outputs, Content appropriateness, Stereotyping | 
 |  | Considerations: | | Awareness of risks and limitations; providing feedback mechanisms to users. | 
 |  | 
| Supported Languages | | da (Danish), sv (Swedish), en (English), no (Norwegian), is (Icelandic) | 
 | 
| Training Details | 
| Data Sources: | | databricks/databricks-dolly-15k, laion/OIG, OpenAssistant/oasst1 | 
 |  | Data Volume: |  |  | Methodology: | | Pretrained using causal language modeling with NeMo Megatron GPT implementation. The instruct models were finetuned on instruction data using chat and raw text formats. | 
 |  | Model Architecture: | | Large decoder-only pretrained transformer language models. | 
 |  | 
| Safety Evaluation | 
| Risk Categories: |  |  | Ethical Considerations: | | Potential for generating biased, incorrect, or harmful content; overrepresentation of some viewpoints. | 
 |  | 
| Responsible Ai Considerations | 
| Fairness: | | Potential bias due to diverse or non-diverse training data. | 
 |  | Transparency: | | Increased communication and transparency sought through a modified RAIL license. | 
 |  | Mitigation Strategies: | | Encouragement for open communication and feedback collection from indirect users. | 
 |  | 
| Input Output |  | 
| Release Notes |  |