Codegemma 2B by google

 ยป  All LLMs  ยป  google  ยป  Codegemma 2B   URL Share it on

  Autotrain compatible   Endpoints compatible   Gemma   Region:us   Safetensors   Sharded   Tensorflow
Model Card on HF ๐Ÿค—: https://huggingface.co/google/codegemma-2b 

Codegemma 2B Benchmarks

Codegemma 2B (google/codegemma-2b)

Codegemma 2B Parameters and Internals

Model Type 
text-to-text, text-to-code
Use Cases 
Areas:
Generative AI, Code-related tasks
Applications:
IDE extension for code, Interactive code learning, Code conversation
Primary Use Cases:
Code completion, Code generation, Instruction following
Limitations:
Intrinsic LLM limitations
Considerations:
Refer to the Gemma model card for evaluation results.
Additional Notes 
N/A
Supported Languages 
English (high proficiency, specifically for code-related tasks)
Training Details 
Data Sources:
publicly available code repositories, open source mathematics datasets, synthetically generated code
Data Volume:
500 billion tokens
Methodology:
FIM tasks with 80% FIM rate, 50-50 PSM/SPM mode
Hardware Used:
TPUv5e
Model Architecture:
Decoder-only
Safety Evaluation 
Methodologies:
Structured evaluations, Internal red-teaming testing
Findings:
Acceptable thresholds for child safety, content safety, representational harms, memorization, large-scale harms
Risk Categories:
Representational harms, Content safety
Ethical Considerations:
Deferred to Gemma model card details
Responsible Ai Considerations 
Fairness:
Evaluated through human evaluation on prompts
Transparency:
Detailed in Gemma model card
Accountability:
Undisclosed
Mitigation Strategies:
Various evaluations and policy adherence as seen in Gemma model details
Input Output 
Input Format:
code prefix/suffix for pretrained, text for instruction-tuned
Accepted Modalities:
text, code
Output Format:
code completion or generation
Performance Tips:
Avoid extra spaces around tokens for completion.
Release Notes 
Version:
2B
Notes:
Fast code completion variant.
Version:
7B
Notes:
Specialized in code completion and generation.
Version:
7B-IT
Notes:
Instruction tuned for chat and instruction-following.
LLM NameCodegemma 2B
Repository ๐Ÿค—https://huggingface.co/google/codegemma-2b 
Model Size2b
Required VRAM5.1 GB
Updated2024-12-21
Maintainergoogle
Model Typegemma
Model Files  5.0 GB: 1-of-2   0.1 GB: 2-of-2
Model ArchitectureGemmaForCausalLM
Licensegemma
Context Length8192
Model Max Length8192
Transformers Version4.39.3
Tokenizer ClassGemmaTokenizer
Padding Token<pad>
Vocabulary Size256000
Torch Data Typebfloat16

Quantized Models of the Codegemma 2B

Model
Likes
Downloads
VRAM
... Codegemma 2B AWQ 4bit Smashed012273 GB
Codegemma 2B GGUF01131 GB

Best Alternatives to Codegemma 2B

Best Alternatives
Context / RAM
Downloads
Likes
Gemma 1.1 2B It8K / 5.1 GB94636152
Gemma Ko 1.1 2B It8K / 5.1 GB48291
Octopus V28K / 5.1 GB490868
EMO 2B8K / 5.1 GB44251
Gemma 2B Ko V08K / 5 GB25270
Gemma2b Lungcancerqa8K / 3.1 GB762
Gemma 2B Ko Dev Pbmt1928K / 5 GB25201
Gemma 2B Data Std8K / 5.1 GB25301
Geko28K / 5.1 GB160
2B Or Not 2B8K / 5.1 GB7525
Note: green Score (e.g. "73.2") means that the model is better than google/codegemma-2b.

Rank the Codegemma 2B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40013 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217