Model Type | |
Use Cases |
Areas: | |
Applications: | |
Primary Use Cases: | English text and code generation |
|
Limitations: | Out-of-scope for languages other than English without compliance, Risk of misuse if violated acceptable use policy |
|
Considerations: | Developers should tune model for safety based on specific applications. |
|
|
Additional Notes | Focused on safe and inclusive text generation practices. Special measures for sensitive applications implemented. |
|
Supported Languages | |
Training Details |
Data Sources: | publicly available online data |
|
Data Volume: | 15T+ tokens pretraining, 10M human-annotated examples fine-tuning |
|
Methodology: | Supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Auto-regressive language model with transformer architecture |
|
|
Safety Evaluation |
Methodologies: | red teaming, adversarial evaluations |
|
Findings: | Residual risks remain; improved model helpfulness and reduced false refusals compared to Llama 2. |
|
Risk Categories: | child safety, cybersecurity |
|
Ethical Considerations: | Residual risks highlighted |
|
|
Responsible Ai Considerations |
Fairness: | Emphasis on inclusivity and openness |
|
Transparency: | Open source license and transparency on safety standards |
|
Accountability: | Meta and developers accountable for use adhering to license terms |
|
Mitigation Strategies: | Llama Guard 2 and Code Shield tools for safety |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use Reinforcement Learning with Human Feedback for optimal tuning. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release with enhanced helpfulness and safety measures. |
|
|
|