Model Type | text-to-text, decoder-only, large language model |
|
Use Cases |
Areas: | Research, Commercial applications |
|
Applications: | text generation, chatbots, text summarization, language learning tools, knowledge exploration |
|
Primary Use Cases: | Content Creation and Communication, Research and Education |
|
Limitations: | Training Data, Context and Task Complexity, Language Ambiguity and Nuance, Factual Accuracy, Common Sense |
|
|
Supported Languages | |
Training Details |
Data Sources: | Web Documents, Code, Mathematics |
|
Data Volume: | 13 trillion tokens (27B model); 8 trillion tokens (9B model) |
|
Methodology: | Built using the same research and technology as Gemini models |
|
Hardware Used: | |
|
Safety Evaluation |
Methodologies: | structured evaluations, internal red-teaming |
|
Risk Categories: | child safety, content safety, representational harms, memorization, large-scale harms |
|
|
Responsible Ai Considerations |
Fairness: | These models underwent careful scrutiny, input data pre-processing described and posterior evaluations reported in this card. |
|
Transparency: | This model card summarizes details on the models' architecture, capabilities, limitations, and evaluation processes. |
|
Mitigation Strategies: | Continuous monitoring and exploration of de-biasing techniques; Guidelines for content safety provided |
|
|
Input Output |
Input Format: | Text string (e.g., question, prompt, document to be summarized) |
|
Accepted Modalities: | |
Output Format: | Generated English-language text |
|
|