| Model Type | |
| Use Cases |
| Areas: | | Research, Commercial Applications |
|
| Limitations: | | LLMs hallucinate, make mistakes, and should not be trusted. Use at your own risk! |
|
| Considerations: | | Same limitations as phi-2 and LLMs in general. |
|
|
| Additional Notes | | Training was supported by the Flemish Supercomputer Center. |
|
| Supported Languages | |
| Training Details |
| Data Sources: | | BramVanroy/ultrachat_200k_dutch, BramVanroy/no_robots_dutch, BramVanroy/belebele_dutch |
|
| Data Volume: | |
| Methodology: | | SFT-tuning (instruction tuning) |
|
| Training Time: | |
| Hardware Used: | | 4 nodes with 4x A100 80GB GPUs each (16 total) |
|
| Model Architecture: | | Derived from microsoft/phi-2 |
|
|
| Input Output |
| Accepted Modalities: | |
| Output Format: | |
|