Model Type | causal decoder-only, text generation |
|
Use Cases |
Areas: | chat, instruction following |
|
Applications: | personal assistant, language tasks |
|
Primary Use Cases: | text generation, conversational AI |
|
Limitations: | Limited language support beyond primary languages, Bias due to training data |
|
Considerations: | Users should consider ethical implications and add safety measures. |
|
|
Additional Notes | Model available as AWQ, GPTQ, and other quantized versions for diverse hardware |
|
Supported Languages | en (full), fr (full), de (full), es (full), it (limited), pt (limited), pl (limited), nl (limited), ro (limited), cs (limited), sv (limited) |
|
Training Details |
Data Sources: | OpenAssistant/oasst1, ehartford/dolphin, tau/sled, tiiuae/falcon-refinedweb |
|
Data Volume: | |
Methodology: | Supervised finetuning and NTK-YaRN context length extension |
|
Context Length: | |
Hardware Used: | |
Model Architecture: | |
|
Responsible Ai Considerations |
Fairness: | Model carries stereotypes and biases from data, needs user guardrails for fair use. |
|
Transparency: | Transparency about data and methods used provided. |
|
Accountability: | Model outputs accountability lies with the user. |
|
Mitigation Strategies: | Users need to implement precautions and guardrails. |
|
|
Input Output |
Input Format: | Text prompt format with integrated chat tokens |
|
Accepted Modalities: | |
Output Format: | Generated text responding to input context |
|
Performance Tips: | Use recommended libraries (e.g., AutoAWQ) for best performance |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Extended context length; includes NTK-YaRN method for context expansion. |
|
|
|