Model Type | |
Use Cases |
Areas: | |
Applications: | Assistant-like chat, Natural language generation |
|
Primary Use Cases: | English language generation tasks |
|
Limitations: | Use in languages other than English without proper fine-tuning, Use that violates applicable laws |
|
Considerations: | Considerations for tailored safety testing for specific applications |
|
|
Additional Notes | Llama 3 supports only English language generation outright. Developers may fine-tune it to support additional languages. |
|
Training Details |
Data Sources: | Publicly available online data |
|
Data Volume: | |
Methodology: | Supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Hardware Used: | Meta's Research SuperCluster, H100-80GB GPUs |
|
Model Architecture: | Auto-regressive language model, optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | Red teaming, Adversarial evaluations |
|
Findings: | Reduction of residual risks, Model refusal mitigation |
|
Risk Categories: | CBRNE, Cybersecurity, Child Safety |
|
Ethical Considerations: | Ethical guidelines for responsible AI deployment outlined |
|
|
Responsible Ai Considerations |
Transparency: | An open approach to AI for better product safety and innovation |
|
Mitigation Strategies: | Updated Responsible Use Guide, Meta Llama Guard 2, Code Shield |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
|