Model Type | |
Use Cases |
Areas: | Research, Commercial applications |
|
Applications: | General purpose AI systems, Memory/compute constrained environments, Latency bound scenarios |
|
Primary Use Cases: | Language and multimodal research, Building blocks for generative AI powered features |
|
Limitations: | Not designed for all downstream purposes |
|
Considerations: | Developers should consider common limitations and adhere to laws. |
|
|
Additional Notes | Optimized for 4K token context length. Quantum int4 ONNX versions available. |
|
Supported Languages | language (multilingual), proficiency (primary focus on English) |
|
Training Details |
Data Sources: | Publicly available documents, High-quality educational data, Newly created synthetic data |
|
Data Volume: | |
Methodology: | Supervised fine-tuning and Direct Preference Optimization |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Dense decoder-only Transformer |
|
|
Safety Evaluation |
Methodologies: | |
Ethical Considerations: | Developers should mitigate accuracy, safety, and fairness risks |
|
|
Responsible Ai Considerations |
Fairness: | The models may reinforce demeaning or negative stereotypes due to training data biases. |
|
Transparency: | Developers should follow transparency best practices. |
|
Accountability: | Developers are responsible for ensuring compliance with laws and regulations. |
|
Mitigation Strategies: | Train with supervision and optimize for safety guidelines. |
|
|
Input Output |
Input Format: | Text format suited for chat prompts. |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Ensure BOS token is included for more reliable results. |
|
|