scientific research on large language models, especially interpretability research
Limitations:
not intended for deployment, generating harmful or offensive text, not suitable for translation or generating text in other languages, not fine-tuned for downstream contexts
Additional Notes
Pythia-160M-deduped was trained on the Pile after global deduplication; current model released after retraining addressing hyperparameter discrepancies; 154 checkpoints provided per model.
Note: green Score (e.g. "73.2") means that the model is better than EleutherAI/pythia-160m-deduped.
Rank the Pythia 160M Deduped Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.