Model Type | |
Use Cases |
Areas: | |
Applications: | |
Primary Use Cases: | instruction-tuned for assistant-like chat |
|
Limitations: | Only available in English |
|
Considerations: | Developers can fine-tune for other languages if compliance is maintained. |
|
|
Additional Notes | Aimed at openness, inclusivity, and helpfulness while maintaining ethical considerations in AI deployment. |
|
Supported Languages | |
Training Details |
Data Sources: | publicly available online data |
|
Data Volume: | |
Methodology: | Pretrained and instruction tuned with supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Hardware Used: | Meta's Research SuperCluster, third-party cloud compute |
|
Model Architecture: | Auto-regressive language model using an optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | red teaming, adversarial evaluations |
|
Risk Categories: | CBRNE, Cybersecurity, Child Safety |
|
|
Responsible Ai Considerations |
Fairness: | Designed to serve diverse backgrounds and perspectives |
|
Transparency: | Active involvement in safety standardization and transparency |
|
Accountability: | Developers are responsible for ensuring safety in their specific use cases |
|
Mitigation Strategies: | Implemented safety mitigation techniques; Purple Llama tools provided for developers. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use transformers or 'llama3' codebase for best performance. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Release of Meta Llama 3 family in 8B and 70B sizes. |
|
|
|