| Model Type | |
| Use Cases |
| Areas: | | Research, Commercial applications |
|
| Applications: | | Content creation, Text generation, Chatbots and conversational AI, Text summarization |
|
| Primary Use Cases: | | Poems, Scripts, Code, Marketing copy, Email drafts |
|
| Limitations: | | Bias from training data, Performance depends on context length, Language ambiguity, Factual inaccuracies, Common sense reasoning |
|
| Considerations: | | Consider biases, the influence of training data, and the complexity of tasks. |
|
|
| Additional Notes | | RecurrentGemma is faster during inference and requires less memory compared to Gemma models. |
|
| Training Details |
| Hardware Used: | |
| Model Architecture: | | Novel recurrent architecture |
|
|
| Safety Evaluation |
| Methodologies: | | Structured evaluations, Internal red-teaming testing |
|
| Risk Categories: | | Child safety, Content safety, Representational harms, Memorization, Large-scale harms |
|
|
| Responsible Ai Considerations |
| Fairness: | | The models underwent careful scrutiny, input data pre-processing described and posterior evaluations reported in this card. |
|
| Transparency: | | Details on models' architecture, capabilities, limitations, and evaluation processes shared. |
|
| Accountability: | | Accountable through summarizing details in the model card. |
|
| Mitigation Strategies: | | Continuous monitoring, exploration of de-biasing techniques, and education on responsible use are encouraged. |
|
|
| Input Output |
| Input Format: | | Text string (e.g., a question, a prompt, or a document to be summarized). |
|
| Accepted Modalities: | |
| Output Format: | | Generated English-language text in response to the input. |
|
|