Model Type | auto-regressive, language model |
|
Use Cases |
Areas: | |
Applications: | question answering, natural language understanding, reading comprehension |
|
Primary Use Cases: | research on large language models, evaluation and mitigation of biases, developing improvement techniques |
|
Limitations: | further risk evaluation required, not trained with human feedback, may generate harmful content |
|
|
Additional Notes | Instruction tuned, converted to int4 via GPTQ method. |
|
Supported Languages | en (excellent), fr (good), es (good), de (good), ru (average), zh (average) |
|
Training Details |
Data Sources: | CCNet, C4, GitHub, Wikipedia, Books, ArXiv, Stack Exchange |
|
Data Volume: | |
Training Time: | December 2022 - February 2023 |
|
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | |
Risk Categories: | gender, religion, race/Color, sexual orientation, age, nationality, disability, physical appearance, socioeconomic status |
|
Ethical Considerations: | Data collected mostly from the Web, contains offensive, harmful, and biased content. |
|
|
Responsible Ai Considerations |
Fairness: | Bias evaluation using RAI datasets for different categories like gender, religion, race, etc. |
|
Transparency: | Data filtered using Kneser-Ney language model and fastText linear classifier based on proximity to Wikipedia. |
|
Mitigation Strategies: | Filtered data based on proximity to Wikipedia text. |
|
|
Input Output |
Input Format: | Instruction and response format |
|
Accepted Modalities: | |
Performance Tips: | For deterministic results, turn off sampling; set specific sampler settings for better performance. |
|
|