Model Type | |
Use Cases |
Areas: | |
Applications: | |
Primary Use Cases: | Natural language generation tasks |
|
Limitations: | Use cases not covered extensively in languages other than English. |
|
Considerations: | Developers should ensure the responsible use of models. |
|
|
Additional Notes | Tuned models optimized for dialogue. High carbon footprint during pretraining offset by Meta's sustainability program. Modeled potential relationships between text sequences to predict next items in sequences safely and effectively. |
|
Supported Languages | English (Primary language for intended use) |
|
Training Details |
Data Sources: | Publicly available online data |
|
Data Volume: | |
Methodology: | Uses a mix of publicly available online data. Fine-tuned using Supervised Fine-Tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF). |
|
Context Length: | |
Hardware Used: | A100-80GB (TDP of 350-400W) |
|
Model Architecture: | Optimized transformer architecture |
|
|
Responsible Ai Considerations |
Fairness: | Model may produce inaccurate, biased, or objectionable outputs. |
|
Transparency: | Transparency measures are in place for users. |
|
Accountability: | Developers should perform safety testing tailored to specific applications. |
|
Mitigation Strategies: | Safety testing and tuning recommended by Meta before deployment. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Specific formatting needs for chat versions, including the use of `INST` and `<>` tags, `BOS` and `EOS` tokens, and appropriate whitespace management. |
|
|