| Use Cases |
| Areas: | | General Chat, Structured Output, Agent Cases (Autogen, Memgpt, Functions), Role-playing |
|
| Applications: | | Coding Assistance, LeetCode Problems |
|
| Primary Use Cases: | |
| Limitations: | | May not be ethically aligned without an additional layer |
|
| Considerations: | | Implement alignment layer before exposing as service. |
|
|
| Additional Notes | | Model is highly compliant to any requests, even unethical ones. It is recommended to implement your own alignment layer. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | ehartford/dolphin, jondurbin/airoboros-2.2.1, ehartford/dolphin-coder, teknium/openhermes, ise-uiuc/Magicoder-OSS-Instruct-75K, ise-uiuc/Magicoder-Evol-Instruct-110K, LDJnr/Capybara |
|
| Methodology: | |
| Context Length: | |
| Training Time: | | 3 days (1.5 epochs on 4x A100s) |
|
| Hardware Used: | |
|
| Responsible Ai Considerations |
| Fairness: | | Dataset was filtered to remove alignment and bias. |
|
| Accountability: | | User is responsible for content created using the model. |
|
| Mitigation Strategies: | | Recommended to implement own alignment layer before exposing as a service. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Implement own alignment layer for ethical and safe deployment. |
|
|