| Model Type | |
| Use Cases |
| Primary Use Cases: | | Instruction-based coding in Python, based on instructions written in English or Russian |
|
| Limitations: | | Model is not aligned to human preferences for safety, No moderation mechanisms, Trained on code based instruction so may produce problematic outputs without filtering |
|
| Considerations: | | Users should be aware of risks, biases, and limitations of the model. |
|
|
| Additional Notes | | This adapter model was trained using `bitsandbytes` quantization config: 4-bit load, nf4 quant type, float16 compute dtype. |
|
| Supported Languages | | ru (high), en (high), Python (high) |
|
| Training Details |
| Data Sources: | | zelkame/ru-stackoverflow-py, MexIvanov/Vezora-Tested-22k-Python-Alpaca-ru, MexIvanov/CodeExercise-Python-27k-ru |
|
| Methodology: | | Special training methods or approaches used, such as fine-tuning techniques. |
|
| Model Architecture: | | A LoRA (Peft) adapter model trained on a mix of publicly available data and machine-translated synthetic python coding datasets. |
|
|
| Responsible Ai Considerations |
| Mitigation Strategies: | | Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. |
|
|
| Input Output |
| Input Format: | | <|system|> ~~ <|user|> {prompt}~~ <|assistant|> |
|
| Accepted Modalities: | |
|