| Model Type | | Transformer-based Language Model, Causal Language Modeling |
|
| Use Cases |
| Areas: | | Research, Scientific Experiments |
|
| Applications: | | Interpretability Research |
|
| Primary Use Cases: | | Analyzing behavior and functionality of large language models |
|
| Limitations: | | Not suitable for translation or non-English text generation, Not intended for deployment in human-facing interactions |
|
| Considerations: | | Text generated may be socially unacceptable or undesirable. Users should conduct risk assessments. |
|
|
| Additional Notes | | Model checkpoints are available on Hugging Face hosted as branches for further fine-tuning. |
|
| Supported Languages | | languages (.English), proficiency (.High) |
|
| Training Details |
| Data Sources: | | The Pile (globally deduplicated) |
|
| Data Volume: | |
| Model Architecture: | |
|
| Safety Evaluation |
| Methodologies: | |
| Risk Categories: | |
| Ethical Considerations: | | The model is trained on the Pile, which is known to contain profanity and offensive text. |
|
|
| Responsible Ai Considerations |
| Fairness: | | The Pile contains biases related to gender, religion, and race. Users should conduct their own risk and bias assessments before deployment. |
|
| Accountability: | | EleutherAI is responsible for the training and release of the model. |
|
| Mitigation Strategies: | | None provided directly; users are advised to curate model outputs before presentation. |
|
|
| Input Output |
| Input Format: | | Text input for causal language modeling. |
|
| Accepted Modalities: | |
| Output Format: | | Text generation as the next token prediction. |
|
| Performance Tips: | | Fine-tune appropriately; ensure model outputs are curated before use. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Renaming of models, retrained with uniform batch sizes and checkpoints. |
|
| Version: | |
| Notes: | | Initial release of models with hyperparameter discrepancies. |
|
|
|