| Model Type | | text generation, decoder-only large language models |
|
| Use Cases |
| Areas: | | content creation, communication, research and education |
|
| Applications: | | text generation, chatbots, text summarization, NLP research, language learning tools |
|
| Primary Use Cases: | | creative text generation, interactive applications, research assistance |
|
| Limitations: | | training data biases, context and task complexity, language ambiguity, factual accuracy, common sense reasoning |
|
| Considerations: | | Guidelines for responsible use as per Responsible Generative AI Toolkit. |
|
|
| Supported Languages | |
| Training Details |
| Data Sources: | | Web Documents, Code, Mathematics |
|
| Data Volume: | | 27B model: 13 trillion tokens |
|
| Methodology: | | Text-to-text, instruction-tuned |
|
| Hardware Used: | |
| Model Architecture: | |
|
| Safety Evaluation |
| Methodologies: | | structured evaluations, internal red-teaming |
|
| Findings: | | acceptable thresholds for internal policies |
|
| Risk Categories: | | child safety, content safety, representational harms, memorization, large-scale harms |
|
| Ethical Considerations: | | This model card summarizes details on the models' architecture, limitations, and evaluation processes. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Careful scrutiny, input data pre-processing described and posterior evaluations reported. |
|
| Transparency: | | This model card summarizes details on the models' architecture, capabilities, limitations, and evaluation processes. |
|
| Accountability: | | Exploration of de-biasing techniques encouraged. |
|
| Mitigation Strategies: | | Continuous monitoring encouraged. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | | Generated English-language text |
|
|