Model Type | |
Use Cases |
Areas: | |
Applications: | memory/compute constrained environments, latency bound scenarios, strong reasoning capabilities such as code, math, and logic |
|
Primary Use Cases: | language and multimodal models research, generative AI powered features |
|
Limitations: | models are not specifically designed or evaluated for all downstream purposes |
|
Considerations: | Developers should evaluate accuracy, safety, and fairness before using in a specific downstream use case. |
|
|
Supported Languages | |
Training Details |
Data Sources: | publicly available documents, selected high-quality educational data, newly created synthetic data, high quality chat format supervised data |
|
Data Volume: | |
Methodology: | Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Dense decoder-only Transformer model |
|
|
Responsible Ai Considerations |
Fairness: | Models can over- or under-represent groups of people and may reinforce negative stereotypes. |
|
Transparency: | Developers should apply responsible AI best practices. |
|
Accountability: | Developers are responsible for ensuring compliance with relevant laws and regulations. |
|
Mitigation Strategies: | Use available safety classifiers or custom solutions appropriate for the use case. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | Generated text in response to input |
|
|