| Model Type | | text-to-text, decoder-only |
|
| Use Cases |
| Areas: | | Research, Commercial applications |
|
| Applications: | | Content Creation, Chatbots, NLP Research, Text Summarization |
|
| Primary Use Cases: | | Question answering, Summarization, Reasoning |
|
| Limitations: | | Biases or gaps in responses, Challenges with open-ended tasks |
|
| Considerations: | | Guidelines provided for responsible use. |
|
|
| Additional Notes | | Pre-trained variants and instruction-tuned for diverse text generation tasks. |
|
| Supported Languages | | English (Full proficiency) |
|
| Training Details |
| Data Sources: | | Web Documents, Code, Mathematics |
|
| Data Volume: | |
| Methodology: | | Instruction-tuned on UltraChat dataset using QLoRA |
|
| Hardware Used: | | Tensor Processing Unit (TPU), TPUv5e |
|
| Model Architecture: | |
|
| Safety Evaluation |
| Methodologies: | | Red-teaming, Human evaluation on safety policies |
|
| Findings: | | Within acceptable thresholds for meeting internal policies |
|
| Risk Categories: | | Text-to-Text Content Safety, Text-to-Text Representational Harms, Memorization, Large-scale harm |
|
|
| Responsible Ai Considerations |
| Fairness: | | Evaluations against WinoBias and BBQ Dataset for representational harms. |
|
| Transparency: | | This model card summarizes details on the models' architecture, capabilities, limitations, and evaluation processes. |
|
| Accountability: | | Guidelines for responsible use with the model provided. |
|
| Mitigation Strategies: | | Continuous monitoring and de-biasing techniques suggested. |
|
|
| Input Output |
| Input Format: | | Text string (e.g., questions, prompts) |
|
| Accepted Modalities: | |
| Output Format: | | Generated English-language text |
|
| Performance Tips: | | Provide well-defined prompts and sufficient context for complex tasks. |
|
|