Mamba 2.8B Chat No Robots by clibrain

 ยป  All LLMs  ยป  clibrain  ยป  Mamba 2.8B Chat No Robots   URL Share it on

  Arxiv:2203.02155   Conversational Dataset:huggingfaceh4/no robot...   En   Endpoints compatible   Pytorch   Region:us   Tensorboard

Mamba 2.8B Chat No Robots Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mamba 2.8B Chat No Robots (clibrain/mamba-2.8b-chat-no_robots)

Mamba 2.8B Chat No Robots Parameters and Internals

Model Type 
text generation
Additional Notes 
The model is inspired by Mamba state space model architecture with focus on efficient handling of information-dense data.
Supported Languages 
English (en)
Training Details 
Data Sources:
HuggingFaceH4/no_robots
Methodology:
fine-tuning on instruction-following data
Model Architecture:
Structured state space model with hardware-aware design
Input Output 
Input Format:
text-based prompts with chat template
Accepted Modalities:
text
Output Format:
generated text responses
LLM NameMamba 2.8B Chat No Robots
Repository ๐Ÿค—https://huggingface.co/clibrain/mamba-2.8b-chat-no_robots 
Model Size2.8b
Required VRAM5.5 GB
Updated2025-02-22
Maintainerclibrain
Model Files  5.5 GB
Supported Languagesen
Model ArchitectureAutoModel
Licensewtfpl
Tokenizer ClassGPTNeoXTokenizer
Padding Token<|endoftext|>
Vocabulary Size50277

Best Alternatives to Mamba 2.8B Chat No Robots

Best Alternatives
Context / RAM
Downloads
Likes
Mamba 2.8B0K / 11.1 GB11608146
Mamba 2.8B Slimpj0K / 11.1 GB2268123
Synatra Mamba Ko 2.8B0K / 5.8 GB591
Mamba Chat 2.8B0K / 5.5 GB493
Mamba 2.8B Instruct Openhermes0K / 5.5 GB8371
Mamba 2.8B CyberSec0K / 5.5 GB449
Bagel DPO 2.8B V0.20K / 11.1 GB1220
Bagel 2.8B V0.20K / 11.1 GB107
...a Financial Headline Sentiment0K / 5.5 GB51
Ct2fast Pythia 2.8B0K / 5.5 GB51
Note: green Score (e.g. "73.2") means that the model is better than clibrain/mamba-2.8b-chat-no_robots.

Rank the Mamba 2.8B Chat No Robots Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227