Model Type | |
Use Cases |
Areas: | commercial applications, research |
|
Applications: | chatbots, natural language generation tasks |
|
Primary Use Cases: | assistant-like interactions |
|
Limitations: | English-only capability., Not intended for use in legally restricted areas. |
|
Considerations: | Ensure compliance with the provided Acceptable Use Policy. |
|
|
Additional Notes | Technically adept users may modify and adapt quantization with stepwise guidance provided in the repository. |
|
Supported Languages | en (high), other_languages (not listed) |
|
Training Details |
Data Sources: | publicly available sources, human-annotated examples |
|
Data Volume: | |
Methodology: | pretraining and fine-tuning with supervised techniques; uses Grouped-Query Attention for scalability in larger models |
|
Context Length: | |
Training Time: | January 2023 to July 2023 |
|
Hardware Used: | A100-80GB GPUs during pretraining |
|
Model Architecture: | optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | internal safety evaluations, comparison with open-source and proprietary models |
|
Findings: | Llama-2-Chat performs better on safety benchmarks than Llama 1 and comparably to some closed-source models. |
|
Risk Categories: | |
Ethical Considerations: | Developers must ensure safety testing and tuning tailored to specific applications. |
|
|
Responsible Ai Considerations |
Fairness: | The model is only tested in English. |
|
Transparency: | The model's limitations and risk areas are highlighted. |
|
Accountability: | Meta is accountable for the model's outputs. |
|
Mitigation Strategies: | Meta offers a Responsible Use Guide to help developers safely use the model. |
|
|
Input Output |
Input Format: | Input text prompts in provided template form. |
|
Accepted Modalities: | |
Output Format: | Textual generation output. |
|
Performance Tips: | Use proper configuration and hardware acceleration for optimal performance. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Release of Llama 2 fine-tuned model with conversational optimization. |
|
|
|