Model Type | |
Use Cases |
Applications: | General purpose AI systems, Research, Commercial applications |
|
Primary Use Cases: | Memory/compute constrained environments, Latency bound scenarios, Strong reasoning: code, math, and logic |
|
Limitations: | Models may not be suitable for high-risk scenarios without assessments. |
|
Considerations: | Models are not specifically evaluated for all downstream purposes. |
|
|
Additional Notes | The Phi-3 Medium models can run on multiple platforms with optimized configurations through ONNX. |
|
Supported Languages | primary_language (English), other_languages () |
|
Training Details |
Data Sources: | Publicly available documents, High-quality educational data, Code, Newly created synthetic data, High quality chat format supervised data |
|
Data Volume: | |
Methodology: | Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Dense decoder-only Transformer |
|
|
Responsible Ai Considerations |
Fairness: | Models trained primarily on English text which affects other languages and dialects. |
|
Transparency: | Developers are advised to follow transparency best practices. |
|
Accountability: | Developers are responsible for ensuring compliance with relevant laws and regulations. |
|
Mitigation Strategies: | Suggestions for using safety classifiers or custom solutions. |
|
|
Input Output |
Input Format: | Chat format with user and assistant roles. |
|
Accepted Modalities: | |
Output Format: | |
|