| Model Type | | Transformer, text generation, NLP, code | 
 | 
| Use Cases | 
| Areas: |  |  | Applications: | | text generation, code generation | 
 |  | Primary Use Cases: | | poem writing, email drafting, story creation, text summarization, Python code writing | 
 |  | Limitations: | | Potential to generate harmful content, Generate inaccurate code and facts, Unreliable responses to instruction, Limited scope for code | 
 |  | Considerations: | | Users should be cautious and critically evaluate outputs. | 
 |  | 
| Additional Notes | | Phi-1.5-generated text/code should be treated as a starting point. Users should verify API uses manually where uncommon packages are involved. | 
 | 
| Supported Languages |  | 
| Training Details | 
| Data Sources: | | same data sources as phi-1, various NLP synthetic texts | 
 |  | Data Volume: |  |  | Training Time: |  |  | Hardware Used: |  |  | Model Architecture: | | Transformer-based with next-word prediction objective | 
 |  | 
| Responsible Ai Considerations | 
| Transparency: | | The model has not undergone instruction fine-tuning. | 
 |  | Mitigation Strategies: | | Model is intended for research to help develop methods to reduce toxicity directly after pretraining. | 
 |  | 
| Input Output | 
| Input Format: | | QA format, Chat format, Code format | 
 |  | Accepted Modalities: |  |  | Output Format: |  |  | Performance Tips: | | Users should update to `transformers` version 4.37.0 or higher. | 
 |  |