Model Type | text generation, decoder-only large language models |
|
Use Cases |
Areas: | content creation, communication, research and education |
|
Applications: | text generation, chatbots, text summarization, NLP research, language learning tools |
|
Primary Use Cases: | creative text generation, interactive applications, research assistance |
|
Limitations: | training data biases, context and task complexity, language ambiguity, factual accuracy, common sense reasoning |
|
Considerations: | Guidelines for responsible use as per Responsible Generative AI Toolkit. |
|
|
Supported Languages | |
Training Details |
Data Sources: | Web Documents, Code, Mathematics |
|
Data Volume: | 27B model: 13 trillion tokens |
|
Methodology: | Text-to-text, instruction-tuned |
|
Hardware Used: | |
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | structured evaluations, internal red-teaming |
|
Findings: | acceptable thresholds for internal policies |
|
Risk Categories: | child safety, content safety, representational harms, memorization, large-scale harms |
|
Ethical Considerations: | This model card summarizes details on the models' architecture, limitations, and evaluation processes. |
|
|
Responsible Ai Considerations |
Fairness: | Careful scrutiny, input data pre-processing described and posterior evaluations reported. |
|
Transparency: | This model card summarizes details on the models' architecture, capabilities, limitations, and evaluation processes. |
|
Accountability: | Exploration of de-biasing techniques encouraged. |
|
Mitigation Strategies: | Continuous monitoring encouraged. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | Generated English-language text |
|
|