Model Type | Multilingual, Instruction-following, MPT architecture |
|
Use Cases |
Areas: | Research, Commercial Applications |
|
Applications: | Reading Comprehension, Brainstorming, Creative Writing |
|
Primary Use Cases: | Instruction-following tasks |
|
Limitations: | Math problems, Reasoning, Factfulness |
|
Considerations: | Users should be aware of biases and limitations. |
|
|
Additional Notes | The model focuses on transparency regarding data, code, and processes. Finetuning datasets are open and commercially permissible. |
|
Supported Languages | primary (Thai), secondary (English), additional () |
|
Training Details |
Data Sources: | laion/OIG, databricks/databricks-dolly-15k, thaisum, scb_mt_enth_2020, garage-bAInd/Open-Platypus, iapp_wiki_qa_squad, pythainlp/han-instruct-dataset-v1.0, cognitivecomputations/dolphin, Hello-SimpleAI/HC3, Muennighoff/xP3x, openai/summarize_from_feedback |
|
Methodology: | Instruction-finetuned, QLoRA with 4 A100 GPUs |
|
Model Architecture: | |
|