Model Type | text-to-text, decoder-only, large language model |
|
Use Cases |
Areas: | Research, Commercial applications |
|
Applications: | Text Generation, Chatbots, Text Summarization, Research, Language Learning |
|
Primary Use Cases: | Question answering, Summarization, Reasoning |
|
Limitations: | Open-ended, highly complex tasks may be challenging, Lacks deep common sense reasoning |
|
Considerations: | Consider dataset biases and misuse potential. |
|
|
Additional Notes | Encouraged feedback from community. Open model for access to innovative AI. |
|
Supported Languages | |
Training Details |
Data Sources: | Web Documents, Code, Mathematics |
|
Data Volume: | |
Methodology: | Trained using RLHF and instruction-tuned techniques |
|
Hardware Used: | |
Model Architecture: | Large language model with text-to-text and decoder-only architecture. |
|
|
Safety Evaluation |
Methodologies: | Red-teaming, Human Evaluation, Automated Testing |
|
Findings: | Models evaluated for content safety, representational harms, memorization, large-scale harm risks., Within acceptable thresholds for internal policies. |
|
Risk Categories: | Child Safety, Content Safety, Representational Harms, Memorization, Dangerous Capabilities |
|
Ethical Considerations: | Monitored for biases and adjusted to mitigate representation harms. |
|
|
Responsible Ai Considerations |
Fairness: | Monitored biases, using evaluations like WinoBias and BBQ. |
|
Transparency: | Open model details summarised in model card. |
|
Accountability: | Developed and maintained by Google with published guidelines for responsible use. |
|
Mitigation Strategies: | Filtering training data, using responsible AI toolkit guidelines. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | Generated English-language text |
|
Performance Tips: | Longer context generally leads to better outputs. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Update with RLHF method, improvements in quality & factuality. |
|
|
|