| Model Type | |
| Use Cases |
| Areas: | | Academic Research, Non-commercial applications |
|
| Applications: | | Text generation for educational purposes, Creative writing |
|
| Primary Use Cases: | | Language translation, Summarization |
|
| Limitations: | | Not suitable for real-time application, High compute cost |
|
| Considerations: | | Best suited for offline research and use with caution regarding biases. |
|
|
| Additional Notes | | Requires a non-commercial use license from Meta AI. |
|
| Training Details |
| Data Sources: | | Common Crawl, GitHub public repos, Wikipedia |
|
| Data Volume: | |
| Methodology: | | Pretraining using transformer architecture |
|
| Context Length: | |
| Training Time: | | ~35 days using 1024 A100 GPUs |
|
| Hardware Used: | |
| Model Architecture: | | Transformer-based architecture with enhanced efficiency techniques. |
|
|
| Safety Evaluation |
| Methodologies: | | Adversarial testing, Bias detection |
|
| Findings: | | Common NLP biases present, improved safety against adversarial attacks |
|
| Risk Categories: | | misinformation, bias, toxicity |
|
| Ethical Considerations: | | Addressing known NLP biases and potential misuse. |
|
|
| Responsible Ai Considerations |
| Fairness: | | Efforts to reduce AI bias incorporated, though challenges remain. |
|
| Transparency: | | Open sourcing with limited data disclosure for transparency. |
|
| Accountability: | | Meta AI is accountable for model development. |
|
| Mitigation Strategies: | | Ongoing evaluation for bias and adversarial robustness. |
|
|
| Input Output |
| Input Format: | | Textual prompts in natural language. |
|
| Accepted Modalities: | |
| Output Format: | | Generated text in response to input prompts. |
|
| Performance Tips: | | Ensure high-quality input text to improve output accuracy. |
|
|
| Release Notes |
| Version: | |
| Date: | |
| Notes: | | Initial release of LLaMA architecture with 70B parameters. |
|
|
|