Model Type | text generation, dialogue optimization |
|
Use Cases |
Areas: | |
Applications: | Assistant-like chat, natural language generation tasks |
|
Primary Use Cases: | Assistant applications, Dialogue management |
|
Limitations: | Not tested exhaustively across languages, Potential for bias and inaccuracy |
|
Considerations: | Developers should ensure thorough safety testing before deployment. |
|
|
Additional Notes | This model is static and trained until July 2023. Expected future versions to improve safety based on feedback. |
|
Supported Languages | |
Training Details |
Data Sources: | A new mix of publicly available online data |
|
Data Volume: | |
Methodology: | Includes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Hardware Used: | Meta's Research Super Cluster, A100-80GB GPUs, with cumulative 3.3M GPU hours |
|
Model Architecture: | Auto-regressive language model using optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | |
Findings: | Potentially unpredictable outputs, model may produce inaccurate or biased responses |
|
Risk Categories: | Inaccuracy, Bias, Objectionable responses |
|
Ethical Considerations: | Pre-deployment safety testing recommended |
|
|
Responsible Ai Considerations |
Fairness: | Testing for fairness and bias conducted in English |
|
Transparency: | Reports available for potential risks |
|
Accountability: | Users are responsible for testing tailored to specific applications |
|
Mitigation Strategies: | Recommendations to perform safety tuning tailored to specific applications |
|
|
Input Output |
Input Format: | Text format with specific tags and tokens such as [INST] and <> |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Ensure the correct sequence and format of tokens for the best performance. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release for commercial and research use, focusing on dialogue optimization. |
|
|
|