LLM Explorer: A Curated Large Language Model Directory and Analytics  // 

Deepseek Math 7B Instruct by deepseek-ai

What open-source LLMs or SLMs are you in search of? 18870 in total.

  Arxiv:2402.03300   Autotrain compatible   Conversational   Endpoints compatible   Has space   Instruct   License:other   Llama   Pytorch   Region:us   Sharded

Rank the Deepseek Math 7B Instruct Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
Deepseek Math 7B Instruct (deepseek-ai/deepseek-math-7b-instruct)

Quantized Models of the Deepseek Math 7B Instruct

Model
Likes
Downloads
VRAM
Deepseek Math 7B Instruct GGUF11122 GB
...epseek Math 7B Instruct AWQ Q4074 GB

Best Alternatives to Deepseek Math 7B Instruct

Best Alternatives
HF Rank
Context/RAM
Downloads
Likes
AIRIC The Mistral59.9532K / 14.4 GB19313
7B54.868K / 15.5 GB262120
...uga 7B Instruct Pl Lora Unload53.542K / 27 GB25732
Telugu Llama2 7B V0 Instruct52.864K / 13.5 GB101610
Chinese Llama 2 7B52.592K / 27 GB3226298
TowerInstruct 7B V0.152.394K / 27.1 GB764252
Vigogne 2 7B Instruct52.022K / 13.5 GB260423
Blossom V2 Llama2 7B51.714K / 13.5 GB28191
Llama2 7B Hf Instruction Lora51.542K / 13.5 GB16620
Vicuna 7B V1.5 PL Lora Unload51.464K / 13.5 GB25930
Note: green Score (e.g. "73.2") means that the model is better than deepseek-ai/deepseek-math-7b-instruct.

Deepseek Math 7B Instruct Parameters and Internals

LLM NameDeepseek Math 7B Instruct
RepositoryOpen on 🤗 
Model Size7b
Required VRAM13.9 GB
Updated2024-02-29
Maintainerdeepseek-ai
Model Typellama
Instruction-BasedYes
Model Files  10.0 GB: 1-of-2   3.9 GB: 2-of-2
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length4096
Model Max Length4096
Transformers Version4.33.1
Tokenizer ClassLlamaTokenizerFast
Beginning of Sentence Token<|begin▁of▁sentence|>
End of Sentence Token<|end▁of▁sentence|>
Vocabulary Size102400
Initializer Range0.02
Torch Data Typebfloat16
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024022003