Recurrentgemma 2B by google

 ยป  All LLMs  ยป  google  ยป  Recurrentgemma 2B   URL Share it on

  Arxiv:1705.03551   Arxiv:1804.06876   Arxiv:1804.09301   Arxiv:1809.02789   Arxiv:1811.00937   Arxiv:1904.09728   Arxiv:1905.07830   Arxiv:1905.10044   Arxiv:1907.10641   Arxiv:1911.01547   Arxiv:1911.11641   Arxiv:2009.03300   Arxiv:2009.11462   Arxiv:2101.11718   Arxiv:2103.03874   Arxiv:2107.03374   Arxiv:2108.07732   Arxiv:2109.07958   Arxiv:2110.08193   Arxiv:2110.14168   Arxiv:2203.09509   Arxiv:2206.04615   Arxiv:2304.06364   Arxiv:2402.19427   Autotrain compatible   Endpoints compatible   Recurrent gemma   Region:us   Safetensors   Sharded   Tensorflow

Recurrentgemma 2B Benchmarks

Recurrentgemma 2B (google/recurrentgemma-2b)

Recurrentgemma 2B Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
Research, Commercial applications
Applications:
Content creation, Text generation, Chatbots and conversational AI, Text summarization
Primary Use Cases:
Poems, Scripts, Code, Marketing copy, Email drafts
Limitations:
Bias from training data, Performance depends on context length, Language ambiguity, Factual inaccuracies, Common sense reasoning
Considerations:
Consider biases, the influence of training data, and the complexity of tasks.
Additional Notes 
RecurrentGemma is faster during inference and requires less memory compared to Gemma models.
Training Details 
Hardware Used:
TPUv5e clusters
Model Architecture:
Novel recurrent architecture
Safety Evaluation 
Methodologies:
Structured evaluations, Internal red-teaming testing
Risk Categories:
Child safety, Content safety, Representational harms, Memorization, Large-scale harms
Responsible Ai Considerations 
Fairness:
The models underwent careful scrutiny, input data pre-processing described and posterior evaluations reported in this card.
Transparency:
Details on models' architecture, capabilities, limitations, and evaluation processes shared.
Accountability:
Accountable through summarizing details in the model card.
Mitigation Strategies:
Continuous monitoring, exploration of de-biasing techniques, and education on responsible use are encouraged.
Input Output 
Input Format:
Text string (e.g., a question, a prompt, or a document to be summarized).
Accepted Modalities:
text
Output Format:
Generated English-language text in response to the input.
LLM NameRecurrentgemma 2B
Repository ๐Ÿค—https://huggingface.co/google/recurrentgemma-2b 
Model Size2b
Required VRAM10.8 GB
Updated2025-01-23
Maintainergoogle
Model Typerecurrent_gemma
Model Files  5.0 GB: 1-of-3   5.0 GB: 2-of-3   0.8 GB: 3-of-3
Model ArchitectureRecurrentGemmaForCausalLM
Licensegemma
Transformers Version4.40.0.dev0
Tokenizer ClassGemmaTokenizer
Padding Token<pad>
Vocabulary Size256000
Torch Data Typefloat32

Best Alternatives to Recurrentgemma 2B

Best Alternatives
Context / RAM
Downloads
Likes
Recurrentgemma 2B It0K / 5.4 GB3976109
...ntgemma 2B It Bnb 4bit Smashed0K / 2.5 GB71
Note: green Score (e.g. "73.2") means that the model is better than google/recurrentgemma-2b.

Rank the Recurrentgemma 2B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 41774 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227