Model Type | text-to-text, decoder-only |
|
Use Cases |
Areas: | Research, Commercial applications |
|
Applications: | Content Creation, Chatbots, NLP Research, Text Summarization |
|
Primary Use Cases: | Question answering, Summarization, Reasoning |
|
Limitations: | Biases or gaps in responses, Challenges with open-ended tasks |
|
Considerations: | Guidelines provided for responsible use. |
|
|
Additional Notes | Pre-trained variants and instruction-tuned for diverse text generation tasks. |
|
Supported Languages | English (Full proficiency) |
|
Training Details |
Data Sources: | Web Documents, Code, Mathematics |
|
Data Volume: | |
Methodology: | Instruction-tuned on UltraChat dataset using QLoRA |
|
Hardware Used: | Tensor Processing Unit (TPU), TPUv5e |
|
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | Red-teaming, Human evaluation on safety policies |
|
Findings: | Within acceptable thresholds for meeting internal policies |
|
Risk Categories: | Text-to-Text Content Safety, Text-to-Text Representational Harms, Memorization, Large-scale harm |
|
|
Responsible Ai Considerations |
Fairness: | Evaluations against WinoBias and BBQ Dataset for representational harms. |
|
Transparency: | This model card summarizes details on the models' architecture, capabilities, limitations, and evaluation processes. |
|
Accountability: | Guidelines for responsible use with the model provided. |
|
Mitigation Strategies: | Continuous monitoring and de-biasing techniques suggested. |
|
|
Input Output |
Input Format: | Text string (e.g., questions, prompts) |
|
Accepted Modalities: | |
Output Format: | Generated English-language text |
|
Performance Tips: | Provide well-defined prompts and sufficient context for complex tasks. |
|
|