Model Type | text generation, instruction-tuned |
|
Use Cases |
Areas: | |
Applications: | |
Primary Use Cases: | natural language generation tasks |
|
Limitations: | English only; out-of-scope for illegal or prohibited use cases |
|
Considerations: | Developers may fine-tune for languages beyond English adhering to specific policies. |
|
|
Additional Notes | The model is intended for English-language applications and could be adapted for other languages. |
|
Supported Languages | languages_supported (English), proficiency () |
|
Training Details |
Data Sources: | |
Data Volume: | |
Methodology: | Supervised fine-tuning, reinforcement learning with human feedback (RLHF) |
|
Context Length: | |
Hardware Used: | Crusoe Energy high performance L40S cluster |
|
Model Architecture: | Transformer with improved RoPE theta, NTK-aware interpolation |
|
|
Safety Evaluation |
Methodologies: | red teaming, adversarial evaluations |
|
Risk Categories: | misinformation, bias, cybersecurity, child safety |
|
Ethical Considerations: | Iterative testing for CBRNE threats |
|
|
Responsible Ai Considerations |
Fairness: | Model is optimized for safety and helpfulness but trade-offs exist. |
|
Transparency: | Steps for safety best practices outlined in Responsible Use Guide. |
|
Accountability: | |
Mitigation Strategies: | Meta Llama Guard 2 and Code Shield provided for safety tailored applications |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
|
Release Notes |
Version: | |
Date: | |
Notes: | Further fine-tuned for assistant-like chat ability; extended context length. |
|
|
|