| Model Type | | text generation, summarization, question answering, reasoning | 
 | 
| Use Cases | 
| Areas: | | content creation, communication, research, education | 
 |  | Applications: | | text generation, chatbots, conversational AI, text summarization | 
 |  | Primary Use Cases: | | question answering, summarization, reasoning | 
 |  | Limitations: | | Influenced by training data biases, Challenges with open-ended or complex tasks | 
 |  | Considerations: | | Model performance impacted by prompt clarity and context length. | 
 |  | 
| Supported Languages |  | 
| Training Details | 
| Data Sources: | | Gemma model family data sources | 
 |  | Methodology: | | Recurrent architecture with pre-training and instruction-tuning. | 
 |  | Hardware Used: |  |  | Model Architecture: |  |  | 
| Safety Evaluation | 
| Methodologies: | | internal red-teaming, structured evaluations | 
 |  | Risk Categories: | | text-to-text content safety, representational harms, memorization, large-scale harm | 
 |  | Ethical Considerations: | | The model adheres to Google's internal safety policies. | 
 |  | 
| Responsible Ai Considerations | 
| Fairness: | | Evaluated against benchmarks like WinoBias and BBQ Dataset for representational harms. | 
 |  | Transparency: | | Details provided in the model card and evaluation processes. | 
 |  | Accountability: | | Accountability not explicitly mentioned. | 
 |  | Mitigation Strategies: | | Provides content safety mechanisms and guidelines. | 
 |  | 
| Input Output | 
| Input Format: | | Text string (question, prompt, document) | 
 |  | Accepted Modalities: |  |  | Output Format: | | Generated English-language text | 
 |  |