| Model Type | | text-to-text, decoder-only, large language model | 
 | 
| Use Cases | 
| Areas: | | Content Creation and Communication, Research and Education | 
 |  | Applications: | | Text Generation, Chatbots and Conversational AI, Text Summarization, NLP Research, Language Learning Tools, Knowledge Exploration | 
 |  | Primary Use Cases: | | Generate creative text formats, Power conversational interfaces, Generate concise summaries | 
 |  | Limitations: | | Training Data Influences, Context and Task Complexity, Language Ambiguity and Nuance, Factual Accuracy Limitations, Common Sense Limitations | 
 |  | Considerations: | | LLMs are better at tasks that can be framed with clear prompts and instructions. Factual accuracy should be verified as LLMs are not knowledge bases. | 
 |  | 
| Additional Notes | | Gemma models are designed for responsible AI development. They include open weights for pre-trained and instruction-tuned variants, enabling wide accessibility and innovation. | 
 | 
| Supported Languages |  | 
| Training Details | 
| Data Sources: | | Web Documents, Code, Mathematics | 
 |  | Data Volume: | | 8 trillion tokens for 9B model | 
 |  | Hardware Used: |  |  | Model Architecture: | | Gemma models are built from the same research and technology used to create the Gemini models. | 
 |  | 
| Safety Evaluation | 
| Methodologies: | | Red-teaming, Benchmark testing | 
 |  | Risk Categories: | | Text-to-Text Content Safety, Text-to-Text Representational Harms, Memorization, Large-scale harm | 
 |  | Ethical Considerations: | | Models were evaluated for child safety, content safety, representational harms, memorization, large-scale harms. | 
 |  | 
| Responsible Ai Considerations | 
| Fairness: | | Models underwent careful scrutiny, input data pre-processing, and posterior evaluations to address socio-cultural biases. | 
 |  | Transparency: | | Details on the models' architecture, capabilities, limitations, and evaluation processes are summarized in the model card. | 
 |  | Accountability: | | Responsible use guidelines are provided, see the Responsible Generative AI Toolkit. | 
 |  | Mitigation Strategies: | | Continuous monitoring and exploration of de-biasing techniques during model training, fine-tuning, and other use cases are encouraged for mitigating perpetuation of biases. | 
 |  | 
| Input Output | 
| Input Format: | | Text string, such as a question, a prompt, or a document to be summarized. | 
 |  | Accepted Modalities: |  |  | Output Format: | | Generated English-language text in response to the input. | 
 |  | Performance Tips: | | Models perform better with clear prompts and sufficient context. | 
 |  |