Training Details |
Data Sources: | OpenAssistant/oasst1 - cleaned dataset, similar to Guanaco, synthetic jokes generation and explanation derived from reddit jokes dataset, synthetic prose generation and rewriting self-chat, Q&A; based on provided context, self instruct augmented logic_inference_oa, de-duped pygmalion dataset, filtered down to RP data, cleaned, english only, 25%, riddle_sense - instruct augmented, hellaswag, updated for detailed explanations w 30K+ rows, gsm8k - instruct augmented, ewof/code-alpaca-instruct-unfiltered synthetic self chat dataset derived from about 1000 rows, subset of QingyiSi/Alpaca-CoT for roleplay and CoT, GPTeacher-General-Instruct, ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, hellaswag - 5K row subset of instruct augmented for concise responses, metaeval/ScienceQA_text_only - instruct for concise responses, openai/summarize_from_feedback - instruct augmented tl;dr summarization |
|
Training Time: | |
Hardware Used: | |
|