Model Type | |
Use Cases |
Areas: | |
Applications: | |
Primary Use Cases: | instruction-tuned models for dialogue |
|
Limitations: | Use must comply with laws and Llama 3 license policy |
|
Considerations: | Developers should perform safety testing and tuning prior to deployment. |
|
|
Additional Notes | Models are optimized for helpfulness and safety through RLHF and SFT. |
|
Supported Languages | |
Training Details |
Data Sources: | publicly available online data |
|
Data Volume: | |
Methodology: | pre-trained, instruction-tuned, RLHF |
|
Context Length: | |
Hardware Used: | Meta's Research SuperCluster, H100-80GB GPUs |
|
Model Architecture: | auto-regressive transformer with Grouped-Query Attention |
|
|
Safety Evaluation |
Methodologies: | red teaming, adversarial evaluations |
|
Findings: | reduced residual risk via safety mitigations |
|
Risk Categories: | misinformation, cybersecurity, child safety |
|
Ethical Considerations: | developers must assess risks for specific use cases |
|
|
Responsible Ai Considerations |
Fairness: | Safety benchmarks are transparent and rigorous. |
|
Transparency: | Evaluations and benchmarks are publicly accessible. |
|
Accountability: | Users and developers must adhere to guidelines and policies. |
|
Mitigation Strategies: | Incorporate safeguards like Meta Llama Guard 2 and Code Shield |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use with transformers pipeline or llama3 codebase for best results. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release of Llama 3 models, including optimized transformers architecture. |
|
|
|