| Model Type | | text-to-text, decoder-only, language model |
|
| Use Cases |
| Areas: | | Research, Commercial applications |
|
| Applications: | | Text Generation, Chatbots and Conversational AI, Text Summarization |
|
| Primary Use Cases: | | Content Creation and Communication, Research and Education |
|
| Limitations: | | Biases in training data, Context complexity, Language Ambiguity, Factual inaccuracies |
|
| Considerations: | | Adherence to privacy regulations, using caution in deployments. |
|
|
| Additional Notes | | Supports various precisions including bfloat16, float16, and float32 for diverse hardware compatibility. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | Web Documents, Code, Mathematics |
|
| Data Volume: | |
| Methodology: | |
| Hardware Used: | |
| Model Architecture: | | Open large language model, text-to-text, decoder-only |
|
|
| Safety Evaluation |
| Methodologies: | | Red-teaming, Human evaluation, Automated evaluation |
|
| Findings: | | Results within acceptable thresholds for child safety, content safety, representational harms, Well known safety benchmarks results provided |
|
| Risk Categories: | | Text-to-Text Content Safety, Text-to-Text Representational Harms, Large-scale harm |
|
| Ethical Considerations: | | Memorization, large-scale harms |
|
|
| Responsible Ai Considerations |
| Fairness: | | Careful scrutiny, input data pre-processing and posterior evaluations done. |
|
| Transparency: | | This model card contains detailed architecture, capabilities, and evaluation processes. |
|
| Accountability: | | Google takes responsibility for releasing the Gemma model. |
|
| Mitigation Strategies: | | Filtering sensitive data, providing guidelines for responsible usage, continuous monitoring and de-biasing. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Utilize appropriate hardware and precision settings for optimal performance. |
|
|
| Release Notes |
| Version: | |
| Notes: | | Gemma 1.1 was trained using a novel RLHF method, addressing aspects like response quality, instruction following, etc. Fixed multi-turn conversation bug, model improvements over previous releases. |
|
|
|