| Model Type | | instruction-tuned, text-generation, chat |
|
| Use Cases |
| Areas: | | research, commercial applications |
|
| Applications: | | text generation, instruction-based chat |
|
| Limitations: | | work in progress, weights may change, model sensitive to generation parameters |
|
|
| Additional Notes | | The model has been fine-tuned on 4096 context length, and it is the first iteration of instruction-tuning NorMistral models. |
|
| Supported Languages | | Norwegian (fluent), Bokmรฅl (fluent), Nynorsk (fluent) |
|
| Training Details |
| Data Sources: | | Aya, OASST 1, OASST 2, OIG-small-chip2, No Robots, Dolly, Glaive code assistant |
|
| Methodology: | | SFT phase without any preference optimization |
|
| Context Length: | |
| Model Architecture: | |
|
| Input Output |
| Input Format: | |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Important to set correct generation parameters (e.g., top_k, top_p, temperature) |
|
|