Model Type | |
Use Cases |
Areas: | research, commercial applications, personal use |
|
Primary Use Cases: | |
Limitations: | May produce hallucinations, Non-determinism in re-generation, Cumulative error potential |
|
Considerations: | Adjust generation parameters for diverse responses |
|
|
Additional Notes | Yi is based on Llama architecture but not a derivative; independently trained. |
|
Supported Languages | English (high), Chinese (high) |
|
Training Details |
Data Sources: | multilingual corpus, custom datasets developed by Yi |
|
Data Volume: | |
Methodology: | Supervised Fine-Tuning (SFT) for chat models |
|
Context Length: | |
Training Time: | |
Hardware Used: | NVIDIA A800, GPU environment |
|
Model Architecture: | Transformer-based, similar to Llama |
|
|
Responsible Ai Considerations |
Fairness: | |
Transparency: | Open-source distribution under Apache 2.0 |
|
Accountability: | |
Mitigation Strategies: | Uses compliance checking algorithms to maximize data compliance |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use appropriate generation settings (temperature, top_p) for task diversity |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Improved coding, math, reasoning abilities |
|
|
|