Use Cases |
Primary Use Cases: | Instruction, conversational, and coding skills; initial agentic abilities; supports function calling |
|
Limitations: | Highly compliant with any requests, even unethical ones, Uncensored, with filtered dataset for more compliance |
|
Considerations: | Implement own alignment layer before deployment. |
|
|
Additional Notes | This model is based on Dolphin-2.9-Mixtral-8x22b and retains much of the original model's performance. |
|
Supported Languages | |
Training Details |
Data Sources: | |
Methodology: | The model was fine-tuned, targeting all layers, and is an extracted expert using SLERP and a custom script. It uses an uncensored dataset with alignment and bias filtered out. |
|
Context Length: | |
Training Time: | |
Hardware Used: | 8xH100 GPUs provided by Crusoe Cloud |
|
Model Architecture: | Mixtral architecture, not fully converted to dense model to retain performance. |
|
|