Model Type | |
Use Cases |
Areas: | |
Applications: | assistant-like chat, natural language generation tasks |
|
Primary Use Cases: | pretrained: general language generation, tuned: chat and assistance |
|
Limitations: | Out-of-scope uses that violate policies or laws, limited to English |
|
Considerations: | Possible inaccuracies, biases, and objectionable content. |
|
|
Additional Notes | Supports long contexts over 1040K. |
|
Supported Languages | |
Training Details |
Data Sources: | |
Data Volume: | 1.4B tokens total for all stages |
|
Methodology: | NTK-aware interpolation for RoPE theta, progressive training on increasing context lengths |
|
Context Length: | |
Hardware Used: | |
Model Architecture: | auto-regressive transformer with RingAttention |
|
|
Safety Evaluation |
Methodologies: | extensive red teaming, adversarial evaluations |
|
Risk Categories: | CBRNE, Cyber Security, Child Safety |
|
Ethical Considerations: | Iterative testing during model training to assess the safety of responses related to CBRNE threats and other adversarial risks. |
|
|
Responsible Ai Considerations |
Fairness: | Safety benchmark standards transparency, comprehensive safety safeguards. |
|
Transparency: | Open approach to AI with community involvement. |
|
Accountability: | Developers responsible for safety deployment based on use case. |
|
Mitigation Strategies: | Use of Purple Llama solutions, thorough safety guides. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use supervised fine-tuning and reinforcement learning with human feedback for optimal results. |
|
|