Model Type | code generation, text generation |
|
Use Cases |
Areas: | |
Applications: | Code generation, Software development |
|
Primary Use Cases: | Generating code snippets, Filling in function bodies |
|
Limitations: | Not suitable for instruction-based code generation, Can generate inaccurate or non-functional code |
|
Considerations: | Use is intended for generating source code similar to input examples, and attribution may be required for verbatim code generation. |
|
|
Additional Notes | The quantized version allows faster inference with reduced memory, suitable for larger batch processing. |
|
Supported Languages | Java (high), JavaScript (high), Python (high) |
|
Training Details |
Data Sources: | |
Data Volume: | |
Methodology: | Trained using GPT-2 architecture with multi-query attention and Fill-in-the-Middle objective |
|
Training Time: | |
Hardware Used: | |
Model Architecture: | GPT-2 model with multi-query attention and Fill-in-the-Middle objective |
|
|
Safety Evaluation |
Ethical Considerations: | The model may generate code that contains bugs or exploits. The code may also require attribution. |
|
|
Responsible Ai Considerations |
Fairness: | The dataset was filtered for permissive licenses. |
|
Accountability: | Users are responsible for ensuring proper attribution to any generated code that matches verbatim the dataset source code. |
|
|
Input Output |
Input Format: | Source code templates or comments in Python, Java, or JavaScript |
|
Accepted Modalities: | |
Output Format: | Code snippets or function body completion |
|
Performance Tips: | The model performs better with context or template inputs similar to training data. |
|
|