LLM Explorer: A Curated Large Language Model Directory and Analytics  // 

CodeFuse DeepSeek 33B by codefuse-ai

What open-source LLMs or SLMs are you in search of? 18870 in total.

 »  All LLMs  »  codefuse-ai  »  CodeFuse DeepSeek 33B   URL Share it on

  Autotrain compatible   Conversational   Endpoints compatible   Has space   License:other   Llama   Pytorch   Region:us   Sharded

CodeFuse DeepSeek 33B Benchmarks

Rank the CodeFuse DeepSeek 33B Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
CodeFuse DeepSeek 33B (codefuse-ai/CodeFuse-DeepSeek-33B)

Quantized Models of the CodeFuse DeepSeek 33B

Model
Likes
Downloads
VRAM
CodeFuse DeepSeek 33B 4bits22438 GB

Best Alternatives to CodeFuse DeepSeek 33B

Best Alternatives
HF Rank
Context/RAM
Downloads
Likes
Zephyrus L1 33B60.612K / 50.6 GB24124
...t 6 Llama 33B Xor MERGED 16bit60.452K / 65.2 GB24012
WizardLM 33B V1.0 Uncensored59.992K / 65.2 GB206658
Fin Llama 33B Merged59.332K / 65.2 GB213814
Samantha 1.1 Llama 33B58.982K / 65.2 GB107016
Chronoboros 33B58.922K / 65.1 GB26359
Scarlett 33B58.812K / 65.3 GB219223
Airochronos 33B58.752K / 65.1 GB26326
Vicuna 33B V1.358.542K / 65.2 GB8395273
Uncensored Frank 33B58.382K / 65.3 GB25585
Note: green Score (e.g. "73.2") means that the model is better than codefuse-ai/CodeFuse-DeepSeek-33B.

CodeFuse DeepSeek 33B Parameters and Internals

LLM NameCodeFuse DeepSeek 33B
RepositoryOpen on 🤗 
Model Size33b
Required VRAM66.5 GB
Updated2024-02-29
Maintainercodefuse-ai
Model Typellama
Model Files  9.7 GB: 1-of-7   9.9 GB: 2-of-7   9.9 GB: 3-of-7   9.8 GB: 4-of-7   9.9 GB: 5-of-7   9.9 GB: 6-of-7   7.4 GB: 7-of-7
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length16384
Model Max Length16384
Transformers Version4.34.1
Tokenizer ClassLlamaTokenizerFast
Beginning of Sentence Token<|begin▁of▁sentence|>
End of Sentence Token<|end▁of▁sentence|>
Vocabulary Size32256
Initializer Range0.02
Torch Data Typebfloat16
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024022003