Research on behavior, functionality, limitations of large language models
Primary Use Cases:
Controlled scientific experiments
Limitations:
Not suitable for human-facing interactions, English language-only models, unsuitable for generating text in other languages, Not fine-tuned for genre prose or commercial chatbots
Considerations:
Conduct risk and bias assessment if fine-tuning; evaluate risks before deployment
Additional Notes
Pythia model suite renamed in January 2023 for clarity
Supported Languages
English (Native)
Training Details
Data Sources:
The Pile, 22 diverse sources including arXiv, CommonCrawl, Project Gutenberg, YouTube subtitles, GitHub
Data Volume:
299,892,736,000 tokens
Model Architecture:
GPT-NeoX
Responsible Ai Considerations
Fairness:
Documented biases with regards to gender, religion, and race (as per Pile paper).
Input Output
Input Format:
String of text for next token prediction.
Accepted Modalities:
text
Output Format:
String (one token at a time)
Release Notes
Version:
Current Release
Date:
January 2023
Notes:
Pyhtia-160M retrained to address hyperparameter discrepancies
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.