| Model Type | | mixture of experts, mixtral |
|
| Use Cases |
| Areas: | | commercial applications, customized LLMs for business |
|
| Limitations: | | Possibility of inappropriate content slipping through, cannot guarantee consistently appropriate behavior |
|
|
| Additional Notes | | Training involved augmenting German data to improve grammatical and syntactical correctness. |
|
| Supported Languages | | English (fluent), German (fluent), French (fluent), Italian (fluent), Spanish (fluent) |
|
| Training Details |
| Data Sources: | | argilla/distilabel-math-preference-dpo, translated Parts of the HuggingFaceH4/ultrafeedback_binarized, Sauerkraut-7b-HerO, German SauerkrautLM-DPO dataset |
|
| Methodology: | |
| Model Architecture: | |
|
| Safety Evaluation |
| Ethical Considerations: | | Despite data cleansing efforts, the possibility of uncensored content slipping through cannot be ruled out. |
|
|
| Responsible Ai Considerations |
| Accountability: | |
| Mitigation Strategies: | | Data cleansing to avoid uncensored content |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
|