Model Type | text generation, instruction tuned |
|
Use Cases |
Areas: | |
Applications: | assistant-like chat, natural language generation tasks |
|
Primary Use Cases: | Dialogue systems, Assistant applications |
|
Limitations: | Use prohibited beyond English and any manner violating applicable laws. |
|
Considerations: | Requires safety testing and tuning for specific applications. |
|
|
Supported Languages | |
Training Details |
Data Sources: | publicly available online data |
|
Data Volume: | |
Methodology: | Supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Hardware Used: | Meta's Research SuperCluster, H100-80GB GPUs |
|
Model Architecture: | Auto-regressive language model with optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | red-teaming, adversarial evaluations |
|
Findings: | model carefully optimized for helpfulness and safety |
|
Risk Categories: | |
Ethical Considerations: | Ethical issues are actively addressed during development. |
|
|
Responsible Ai Considerations |
Fairness: | Model openly accessible with community contribution encouraged. |
|
Transparency: | Active contributions to open consortiums like AI Alliance. |
|
Accountability: | Efforts in place to limit misuse and support open community. |
|
Mitigation Strategies: | Added safeguards like Meta Llama Guard 2 and Code Shield to mitigate risks. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
|
Release Notes |
Date: | |
Notes: | Initial release of the instruction tuned models of Llama 3. |
|
|
|