Not suitable for multilingual, Not suitable for code generation
Additional Notes
Training data was filtered with a toxicity classifier. The model is case-sensitive and differentiates between 'finnish' and 'Finnish'. Known for its strong non-arithmetic task performance in 0-shot settings.
Supported Languages
Finnish (Proficient)
Training Details
Data Sources:
Finnish-NLP/CulturaX_fi_cleaned, Finnish-NLP/HPLT_1.2_fi_cleaned, Finnish-NLP/wikipedia_20231101_fi_cleaned, Finnish-NLP/Reddit_fi_2006_2022, Yle Finnish News Archive 2011-2018, Yle Finnish News Archive 2019-2020, Finnish News Agency Archive (STT), The Suomi24 Sentences Corpus, Project LΓΆnnrot, Finnish parliament speeches, intfloat/multilingual_cc_news, fi-news-corpus, Finnish higher education public theses, Finnish single-turn instruction-following datasets
Data Volume:
41 billion tokens
Methodology:
2-stage pretraining with instruction-following examples
Note: green Score (e.g. "73.2") means that the model is better than Finnish-NLP/Ahma-7B.
Rank the Ahma 7B Capabilities
π Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! π
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.