Model Type | |
Use Cases |
Areas: | Academic Research, Non-commercial applications |
|
Applications: | Text generation for educational purposes, Creative writing |
|
Primary Use Cases: | Language translation, Summarization |
|
Limitations: | Not suitable for real-time application, High compute cost |
|
Considerations: | Best suited for offline research and use with caution regarding biases. |
|
|
Additional Notes | Requires a non-commercial use license from Meta AI. |
|
Training Details |
Data Sources: | Common Crawl, GitHub public repos, Wikipedia |
|
Data Volume: | |
Methodology: | Pretraining using transformer architecture |
|
Context Length: | |
Training Time: | ~35 days using 1024 A100 GPUs |
|
Hardware Used: | |
Model Architecture: | Transformer-based architecture with enhanced efficiency techniques. |
|
|
Safety Evaluation |
Methodologies: | Adversarial testing, Bias detection |
|
Findings: | Common NLP biases present, improved safety against adversarial attacks |
|
Risk Categories: | misinformation, bias, toxicity |
|
Ethical Considerations: | Addressing known NLP biases and potential misuse. |
|
|
Responsible Ai Considerations |
Fairness: | Efforts to reduce AI bias incorporated, though challenges remain. |
|
Transparency: | Open sourcing with limited data disclosure for transparency. |
|
Accountability: | Meta AI is accountable for model development. |
|
Mitigation Strategies: | Ongoing evaluation for bias and adversarial robustness. |
|
|
Input Output |
Input Format: | Textual prompts in natural language. |
|
Accepted Modalities: | |
Output Format: | Generated text in response to input prompts. |
|
Performance Tips: | Ensure high-quality input text to improve output accuracy. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release of LLaMA architecture with 70B parameters. |
|
|
|