Model Type | |
Use Cases |
Areas: | |
Applications: | general purpose AI systems |
|
Primary Use Cases: | memory/compute constrained environments, latency bound scenarios, strong reasoning (especially code, math and logic) |
|
Considerations: | Developers should consider common limitations of language models as they select use cases and evaluate and mitigate for accuracy, safety, and fairness. |
|
|
Additional Notes | Activation Aware Quantization (AWQ) works by identifying the top 1% most salient weights that are most important for maintaining accuracy and quantizing the remaining 99% of weights. |
|
Training Details |
Data Sources: | synthetic data, filtered publicly available websites |
|
Methodology: | supervised fine-tuning, proximal policy optimization, direct preference optimization |
|
Context Length: | |
|
Release Notes |
Version: | Phi-3.5-Mini-Instruct ONNX |
|
Notes: | Update over the instruction-tuned Phi-3 Mini ONNX model release. |
|
|
|