| Model Type | | large, decoder-only, transformer, autoregressive |
|
| Use Cases |
| Areas: | | research, evaluation of Large Language Models capabilities |
|
| Primary Use Cases: | | Validating the model and collecting feedback on Large Language Models. |
|
| Limitations: | | Bias, Safety, Generation diversity, Hallucination, Overrepresentation of some viewpoints, Discriminatory language, Inaccurate information generation, Repetitive outputs, Content appropriateness, Stereotyping |
|
| Considerations: | | Awareness of risks and limitations; providing feedback mechanisms to users. |
|
|
| Supported Languages | | da (Danish), sv (Swedish), en (English), no (Norwegian), is (Icelandic) |
|
| Training Details |
| Data Sources: | | databricks/databricks-dolly-15k, laion/OIG, OpenAssistant/oasst1 |
|
| Data Volume: | |
| Methodology: | | Pretrained using causal language modeling with NeMo Megatron GPT implementation. The instruct models were finetuned on instruction data using chat and raw text formats. |
|
| Model Architecture: | | Large decoder-only pretrained transformer language models. |
|
|
| Safety Evaluation |
| Risk Categories: | |
| Ethical Considerations: | | Potential for generating biased, incorrect, or harmful content; overrepresentation of some viewpoints. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Potential bias due to diverse or non-diverse training data. |
|
| Transparency: | | Increased communication and transparency sought through a modified RAIL license. |
|
| Mitigation Strategies: | | Encouragement for open communication and feedback collection from indirect users. |
|
|
| Input Output | |
| Release Notes | |