Model Type | instruction-tuned, text-generation, chat |
|
Use Cases |
Areas: | research, commercial applications |
|
Applications: | text generation, instruction-based chat |
|
Limitations: | work in progress, weights may change, model sensitive to generation parameters |
|
|
Additional Notes | The model has been fine-tuned on 4096 context length, and it is the first iteration of instruction-tuning NorMistral models. |
|
Supported Languages | Norwegian (fluent), Bokmรฅl (fluent), Nynorsk (fluent) |
|
Training Details |
Data Sources: | Aya, OASST 1, OASST 2, OIG-small-chip2, No Robots, Dolly, Glaive code assistant |
|
Methodology: | SFT phase without any preference optimization |
|
Context Length: | |
Model Architecture: | |
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Important to set correct generation parameters (e.g., top_k, top_p, temperature) |
|
|