TinyStoriesChinese 110M is an open-source language model by fzmnm. Features: 110m LLM, VRAM: 0.4GB, Context: 1K, License: cc, LLM Explorer Score: 0.14.
Educational purposes, Machine learning training for beginners
Applications:
Generation of short Chinese stories
Primary Use Cases:
Text generation for educational tools
Limitations:
Consistency and depth of logical capabilities vary
Considerations:
Focused on simple, narrowly defined tasks; ideal for experimentation
Additional Notes
Demonstrates potential of smaller models to produce coherent text without extensive computational resources.
Supported Languages
zh (Proficient)
Training Details
Data Sources:
adam89/TinyStoriesChinese
Data Volume:
2G tokens
Methodology:
Standard llama2 format Transformer with 12 layers, 12 attention heads, and a hidden size of 768. The model utilizes a 1024 context window with Relative Positional Encodings (RoPE) and a vocabulary size of 5,000.
Note: green Score (e.g. "73.2") means that the model is better than fzmnm/TinyStoriesChinese-110M.
Rank the TinyStoriesChinese 110M Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.