Research on large language models, Exploring applications such as question answering and reading comprehension, Evaluating and mitigating biases, Determining capabilities and limitations of models
Limitations:
Base model not suitable for downstream applications without risk evaluation
Supported Languages
en (High proficiency), others (Included 20 languages, mainly supports English)
1T tokens with different breakdowns for different model sizes
Model Architecture:
transformer architecture
Responsible Ai Considerations
Fairness:
Model reflects biases from web sources. Evaluated biases include gender, religion, race, sexual orientation, age, nationality, disability, physical appearance, and socioeconomic status.
Transparency:
Model trained using web-sourced data which may contain biased and harmful content.
Accountability:
Use GitHub repository to raise questions or comments.
Mitigation Strategies:
Filtered data based on proximity to Wikipedia text using a Kneser-Ney language model and fastText linear classifier.
Note: green Score (e.g. "73.2") means that the model is better than heegyu/LIMA-13b.
Rank the LIMA 13B Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52758 in total.