Model Type | |
Use Cases |
Areas: | Research, Commercial applications |
|
Applications: | |
Primary Use Cases: | Natural language generation tasks |
|
Limitations: | May amplify biases contained in the training data, Potential for inaccuracies in generated responses |
|
Considerations: | Developers to address unforeseen misuse and ensure compliance with industry standards. |
|
|
Additional Notes | Continued support will be added to the `transformers` library. |
|
Supported Languages | English (fluent), Multilingual (basic) |
|
Training Details |
Data Sources: | English and multilingual text, code from webpages, dialogue, articles |
|
Data Volume: | |
Methodology: | Pruning with continued training via distillation |
|
Context Length: | |
Training Time: | July 29, 2024 - Aug 3, 2024 |
|
Hardware Used: | NVIDIA hardware (specific microarchitectures mentioned) |
|
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | 5-shot and zero-shot performance evaluation, Code generation performance |
|
Findings: | Shows potential bias and possibility of generating undesirable text |
|
Risk Categories: | bias, toxicity, inaccuracy |
|
Ethical Considerations: | NVIDIA promotes Trustworthy AI and requires adhering to their terms of service |
|
|
Responsible Ai Considerations |
Fairness: | Developers must ensure fair and unbiased results in their applications. |
|
Transparency: | Models should be used with understanding of potential biases. |
|
Accountability: | Users must take responsibility for deploying the model appropriately. |
|
Mitigation Strategies: | NVIDIA encourages use with internal model teams and reporting vulnerabilities. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Optimal performance with prompt lengths within 8k characters. |
|
|