Model Type | instruct fine-tuned model |
|
Use Cases |
Areas: | local intelligence, on-device computing, at-the-edge use cases |
|
Applications: | |
Primary Use Cases: | non-commercial research purposes |
|
Limitations: | Only for non-commercial research purposes |
|
|
Additional Notes | Trained with a 128k context window utilizing interleaved sliding-window attention. |
|
Supported Languages | en (English), fr (French), de (German), es (Spanish), it (Italian), pt (Portuguese), zh (Chinese), ja (Japanese), ru (Russian), ko (Korean) |
|
Training Details |
Context Length: | |
Model Architecture: | |
|
Input Output |
Input Format: | V3-Tekken tokenizer format |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use Mistral Inference or vLLM for optimized performance. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Release of Ministral-8B-Instruct-2410 under Mistral Research License |
|
|
|