Model Type | |
Use Cases |
Areas: | Research, Commercial Applications |
|
Applications: | Assistant-like chat, Natural language generation tasks |
|
Primary Use Cases: | Intended for English dialogue and assistant-like functionalities |
|
Limitations: | Not suitable for legal compliance violations, Testing performed primarily in English |
|
Considerations: | Conduct safety testing tailored to specific applications before deployment. |
|
|
Additional Notes | Pretraining data cut off in Sep 2022; latest tuning data from July 2023. |
|
Supported Languages | |
Training Details |
Data Sources: | A new mix of publicly available online data |
|
Data Volume: | |
Methodology: | Auto regressive transformer with SFT and RLHF |
|
Context Length: | |
Training Time: | Between January 2023 and July 2023 |
|
Hardware Used: | Meta's Research Super Cluster, production clusters for pretraining |
|
Model Architecture: | Optimized transformer architecture |
|
|
Safety Evaluation |
Methodologies: | Supervised fine-tuning, Reinforcement learning with human feedback, Automatic safety benchmarks |
|
Findings: | On par with closed-source models like ChatGPT and PaLM |
|
Risk Categories: | Inaccurate or biased outputs, Other objectionable responses |
|
Ethical Considerations: | Refer to Responsible Use Guide for detailed information. |
|
|
Responsible Ai Considerations |
Fairness: | Testing conducted only in English. |
|
Transparency: | Details provided in accompanying documentation. |
|
Accountability: | Meta oversees the outputs, encourages safety testing before deployment. |
|
Mitigation Strategies: | Future versions will incorporate community feedback for improved safety. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | Models generate text only. |
|
Performance Tips: | Ensure VRAM and software requirements are met for optimal performance. |
|
|
Release Notes |
Version: | |
Notes: | Multiple GPTQ quantization options; optimized for hardware and requirements. |
|
|
|