Model Type | auto-regressive language model, transformer architecture |
|
Use Cases |
Areas: | research on large language models |
|
Primary Use Cases: | question answering, natural language understanding, reading comprehension |
|
Limitations: | generation of misinformation, generation of harmful, biased or offensive content |
|
Considerations: | Should not be used on downstream applications without further investigation and mitigations of risks. |
|
|
Additional Notes | Model date: Trained between December 2022 and February 2023. Model version: 1. |
|
Supported Languages | bg (unknown), ca (unknown), cs (unknown), da (unknown), de (unknown), en (better performance expected), es (unknown), fr (unknown), hr (unknown), hu (unknown), it (unknown), nl (unknown), pl (unknown), pt (unknown), ro (unknown), ru (unknown), sl (unknown), sr (unknown), sv (unknown), uk (unknown) |
|
Training Details |
Data Sources: | CCNet, C4, GitHub, Wikipedia, Books, ArXiv, Stack Exchange |
|
|
Safety Evaluation |
Risk Categories: | |
Ethical Considerations: | Data contains offensive, harmful, and biased content. Evaluated on RAI datasets for biases. |
|
|
Responsible Ai Considerations |
Fairness: | Evaluated on RAI datasets for biases in gender, religion, race, sexual orientation, age, nationality, disability, physical appearance, and socio-economic status. |
|
Transparency: | Results on evaluation datasets and ethical considerations are mentioned. |
|
Mitigation Strategies: | Data filtered based on proximity to Wikipedia text and references using a Kneser-Ney language model and a fastText linear classifier. |
|
|