Model Type | text-to-text, text-to-code |
|
Use Cases |
Areas: | Generative AI, Code-related tasks |
|
Applications: | IDE extension for code, Interactive code learning, Code conversation |
|
Primary Use Cases: | Code completion, Code generation, Instruction following |
|
Limitations: | Intrinsic LLM limitations |
|
Considerations: | Refer to the Gemma model card for evaluation results. |
|
|
Additional Notes | |
Supported Languages | English (high proficiency, specifically for code-related tasks) |
|
Training Details |
Data Sources: | publicly available code repositories, open source mathematics datasets, synthetically generated code |
|
Data Volume: | |
Methodology: | FIM tasks with 80% FIM rate, 50-50 PSM/SPM mode |
|
Hardware Used: | |
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | Structured evaluations, Internal red-teaming testing |
|
Findings: | Acceptable thresholds for child safety, content safety, representational harms, memorization, large-scale harms |
|
Risk Categories: | Representational harms, Content safety |
|
Ethical Considerations: | Deferred to Gemma model card details |
|
|
Responsible Ai Considerations |
Fairness: | Evaluated through human evaluation on prompts |
|
Transparency: | Detailed in Gemma model card |
|
Accountability: | |
Mitigation Strategies: | Various evaluations and policy adherence as seen in Gemma model details |
|
|
Input Output |
Input Format: | code prefix/suffix for pretrained, text for instruction-tuned |
|
Accepted Modalities: | |
Output Format: | code completion or generation |
|
Performance Tips: | Avoid extra spaces around tokens for completion. |
|
|
Release Notes |
Version: | |
Notes: | Fast code completion variant. |
|
Version: | |
Notes: | Specialized in code completion and generation. |
|
Version: | |
Notes: | Instruction tuned for chat and instruction-following. |
|
|
|