| Model Type | |
| Use Cases |
| Areas: | | research, commercial applications, personal use |
|
| Primary Use Cases: | |
| Limitations: | | May produce hallucinations, Non-determinism in re-generation, Cumulative error potential |
|
| Considerations: | | Adjust generation parameters for diverse responses |
|
|
| Additional Notes | | Yi is based on Llama architecture but not a derivative; independently trained. |
|
| Supported Languages | | English (high), Chinese (high) |
|
| Training Details |
| Data Sources: | | multilingual corpus, custom datasets developed by Yi |
|
| Data Volume: | |
| Methodology: | | Supervised Fine-Tuning (SFT) for chat models |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | | NVIDIA A800, GPU environment |
|
| Model Architecture: | | Transformer-based, similar to Llama |
|
|
| Responsible Ai Considerations |
| Fairness: | |
| Transparency: | | Open-source distribution under Apache 2.0 |
|
| Accountability: | |
| Mitigation Strategies: | | Uses compliance checking algorithms to maximize data compliance |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use appropriate generation settings (temperature, top_p) for task diversity |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Improved coding, math, reasoning abilities |
|
|
|