| Model Type | | text-to-text, decoder-only, large language model |
|
| Use Cases |
| Areas: | | content creation, research, communication |
|
| Applications: | | text generation, chatbots, conversational AI, text summarization |
|
| Primary Use Cases: | | question answering, summarization, reasoning |
|
| Limitations: | | bias, factual inaccuracy, common sense issues |
|
| Considerations: | | Developers are encouraged to apply privacy-preserving techniques and adhere to the Responsible Generative AI Toolkit. |
|
|
| Additional Notes | | These models are optimized for performance and responsible AI use, providing accessibility to advanced AI models. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | Web Documents, Code, Mathematics |
|
| Data Volume: | |
| Methodology: | | Training was done using JAX and ML Pathways |
|
| Hardware Used: | |
| Model Architecture: | | not specified in the data |
|
|
| Safety Evaluation |
| Methodologies: | | structured evaluations, internal red-teaming testing |
|
| Findings: | | Within acceptable thresholds for meeting internal policies |
|
| Risk Categories: | | Text-to-Text Content Safety, Text-to-Text Representational Harms, Memorization, Large-scale harm |
|
| Ethical Considerations: | | Focused on safety, fairness, and privacy. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Scrutiny and pre-processing of input data to handle biases. |
|
| Transparency: | |
| Accountability: | | Responsibility lies with the developers using the model. |
|
| Mitigation Strategies: | | Continuous monitoring and the exploration of de-biasing techniques are encouraged. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | | Generated English-language text |
|
|