Model Type | |
Additional Notes | Significant improvements in knowledge, coding, mathematics, and instruction following abilities. Multilingual support for more than 29 languages. Long-context support up to 128K tokens, generating up to 8K tokens. Robust against diverse system prompts, enhancing chatbot capabilities. |
|
Supported Languages | English (Fluent), Chinese (Fluent), French (Fluent), Spanish (Fluent), Portuguese (Fluent), German (Fluent), Italian (Fluent), Russian (Fluent), Japanese (Fluent), Korean (Fluent), Vietnamese (Fluent), Thai (Fluent), Arabic (Fluent) |
|
Training Details |
Context Length: | |
Model Architecture: | Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias and tied word embeddings |
|
|
Input Output |
Input Format: | Structured input using predefined roles in JSON format |
|
Accepted Modalities: | |
Output Format: | |
|