Model Type | mixture of experts, mixtral |
|
Use Cases |
Areas: | commercial applications, customized LLMs for business |
|
Limitations: | Possibility of inappropriate content slipping through, cannot guarantee consistently appropriate behavior |
|
|
Additional Notes | Training involved augmenting German data to improve grammatical and syntactical correctness. |
|
Supported Languages | English (fluent), German (fluent), French (fluent), Italian (fluent), Spanish (fluent) |
|
Training Details |
Data Sources: | argilla/distilabel-math-preference-dpo, translated Parts of the HuggingFaceH4/ultrafeedback_binarized, Sauerkraut-7b-HerO, German SauerkrautLM-DPO dataset |
|
Methodology: | |
Model Architecture: | |
|
Safety Evaluation |
Ethical Considerations: | Despite data cleansing efforts, the possibility of uncensored content slipping through cannot be ruled out. |
|
|
Responsible Ai Considerations |
Accountability: | |
Mitigation Strategies: | Data cleansing to avoid uncensored content |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
|