| Model Type | | text generation, code synthesis |
|
| Use Cases |
| Areas: | |
| Applications: | | Code synthesis, Code understanding, Python specific tasks |
|
| Primary Use Cases: | | Commercial code generation, Research projects in programming and AI |
|
| Limitations: | | Not for non-English languages, Should not violate laws or regulations |
|
| Considerations: | | Adhere to legal compliances. |
|
|
| Additional Notes | | Intended for Python code tasks; requires specific handling for different operations. |
|
| Supported Languages | | English (proficient), Python (proficient) |
|
| Training Details |
| Data Sources: | | Same data as Llama 2 with different weights |
|
| Methodology: | | Fine-tuned with additional instruction data for CodeLlama - Instruct |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | | Metaβs Research Super Cluster, A100-80GB GPUs |
|
| Model Architecture: | | Autoregressive language model using optimized transformer architectures |
|
|
| Responsible Ai Considerations |
| Fairness: | | Tested primarily in English. |
|
| Transparency: | | Model's potential outputs cannot be predicted in advance. |
|
| Accountability: | | Developers should perform safety testing and tuning tailored to their specific applications before deployment. |
|
| Mitigation Strategies: | | Please see the Responsible Use Guide for more details. |
|
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Use 'trust_remote_code=True' for optimal operation due to a change in the RoPE Theta value |
|
|