Model Type | |
Use Cases |
Areas: | |
Applications: | Memory/compute constrained environments, Latency bound scenarios, Strong reasoning tasks like code, math and logic |
|
Primary Use Cases: | language and multimodal research, generative AI-powered features |
|
Limitations: | not specifically evaluated for all downstream purposes, performance varies across different modalities |
|
Considerations: | Developers should apply debiasing and further mitigate for accuracy, safety, and fairness. |
|
|
Supported Languages | languages_supported (en), proficiency_levels () |
|
Training Details |
Data Sources: | Publicly available documents filtered for quality, high-quality educational data, code |
|
Data Volume: | |
Methodology: | Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | 3.8B parameter dense decoder-only Transformer model |
|
|
Safety Evaluation |
Methodologies: | Supervised fine-tuning (SFT), Direct Preference Optimization (DPO) |
|
Findings: | Can potentially behave unfairly or offend, Possibility of generating nonsensical content, Quality of Service may vary based on language variety |
|
Risk Categories: | misinformation, stereotype perpetuation |
|
Ethical Considerations: | Developers must adhere to responsible AI practices and ensure compliance with laws and regulations. |
|
|
Responsible Ai Considerations |
Fairness: | Models may under/over-represent groups and decisions on use-cases should be sensitive to model limitations. |
|
Transparency: | Detailed transparency related to the training and evaluation process is provided. |
|
Accountability: | Developers are responsible for ensuring fair and compliant use. |
|
Mitigation Strategies: | Supervised fine-tuning and direct preference optimizations are used to align with human preferences and safety guidelines. |
|
|
Input Output |
Input Format: | Chat format. E.g. <|user|>Question<|end|><|assistant|>... |
|
Accepted Modalities: | |
Output Format: | Generated text in response to inputs |
|
|