Model Type | text-generation-inference, transformers |
|
Use Cases |
Areas: | |
Applications: | Assistant-like chat, Natural language generation tasks |
|
Primary Use Cases: | Multilingual dialogue, Synthetic data generation, Data distillation |
|
Limitations: | Use in non-supported languages requires additional tuning and responsibility by developers |
|
Considerations: | Refer to Responsible Use Guide for language use beyond the eight supported languages. |
|
|
Additional Notes | Model trained 2x faster with Unsloth and Hugging Face's TRL library. |
|
Supported Languages | English (yes), German (yes), French (yes), Italian (yes), Portuguese (yes), Hindi (yes), Spanish (yes), Thai (yes) |
|
Training Details |
Data Sources: | argilla/distilabel-intel-orca-kto |
|
Data Volume: | |
Methodology: | KTO Fine tuning: Kahneman-Tversky Optimization (KTO) |
|
Context Length: | |
Model Architecture: | Auto-regressive language model with optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | Evaluation with adversarial datasets, Red teaming exercises |
|
Risk Categories: | CBRNE helpfulness, Child Safety, Cyber attack enablement |
|
Ethical Considerations: | Safety testing and tuning should be tailored to specific applications; potential for biased, objectionable, or inaccurate outputs. |
|
|
Responsible Ai Considerations |
Accountability: | Developers are responsible for deploying safeguards for their specific use cases. |
|
Mitigation Strategies: | Following Responsible Use Guide; integration of safeguards like Llama Guard 3, Prompt Guard, and Code Shield. |
|
|
Input Output |
Input Format: | ChatML or Alpaca prompt template |
|
Accepted Modalities: | |
Output Format: | Multilingual Text and code |
|
|