Model Type | |
Use Cases |
Areas: | Research, Commercial Applications |
|
Limitations: | LLMs hallucinate, make mistakes, and should not be trusted. Use at your own risk! |
|
Considerations: | Same limitations as phi-2 and LLMs in general. |
|
|
Additional Notes | Training was supported by the Flemish Supercomputer Center. |
|
Supported Languages | |
Training Details |
Data Sources: | BramVanroy/ultrachat_200k_dutch, BramVanroy/no_robots_dutch, BramVanroy/belebele_dutch |
|
Data Volume: | |
Methodology: | SFT-tuning (instruction tuning) |
|
Training Time: | |
Hardware Used: | 4 nodes with 4x A100 80GB GPUs each (16 total) |
|
Model Architecture: | Derived from microsoft/phi-2 |
|
|
Input Output |
Accepted Modalities: | |
Output Format: | |
|