Model Type | text generation, instruction tuned |
|
Use Cases |
Areas: | |
Applications: | instruction-tuned for chat applications |
|
Primary Use Cases: | chat-oriented generative tasks |
|
Limitations: | Not suitable for language other than English, restricted under Acceptable Use Policy |
|
Considerations: | Developers encouraged to implement safety assessments for specific applications. |
|
|
Additional Notes | Model supports fine-tuning for languages beyond English under compliance with the license and use policy. |
|
Supported Languages | |
Training Details |
Data Sources: | publicly available online data |
|
Data Volume: | |
Methodology: | supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Training Time: | 7.7M GPU hours on H100-80GB |
|
Hardware Used: | |
Model Architecture: | optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | red teaming, adversarial evaluations |
|
Findings: | residual risks remain, model refusals reduced |
|
Risk Categories: | CBRNE, cybersecurity, child safety |
|
Ethical Considerations: | Ethical considerations include avoiding misuse of AI in harmful areas. |
|
|
Responsible Ai Considerations |
Fairness: | Model is optimized to balance helpfulness and alignment, with considerations for avoiding biases. |
|
Transparency: | Open source release with detailed documentation and responsible use guidelines. |
|
Accountability: | Users must comply with license terms and acceptable use policy. |
|
Mitigation Strategies: | Safety tools like Meta Llama Guard and Code Shield provided. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use appropriate hardware and fine-tuning methods for optimal performance. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release of Llama 3 models. |
|
|
|