Mistral 7B Instruct V0.1 GPTQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Mistral 7B Instruct V0.1 GPTQ   URL Share it on

  4-bit   Autotrain compatible Base model:mistralai/mistral-7... Base model:quantized:mistralai...   Conversational   Finetuned   Gptq   Instruct   Mistral   Quantized   Region:us   Safetensors

Mistral 7B Instruct V0.1 GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mistral 7B Instruct V0.1 GPTQ (TheBloke/Mistral-7B-Instruct-v0.1-GPTQ)

Mistral 7B Instruct V0.1 GPTQ Parameters and Internals

Model Type 
mistral
Additional Notes 
It does not have any moderation mechanisms.
Training Details 
Methodology:
instruct fine-tuning using a variety of publicly available conversation datasets.
Model Architecture:
Based on Mistral-7B-v0.1 with architecture choices: Grouped-Query Attention, Sliding-Window Attention, Byte-fallback BPE tokenizer.
Input Output 
Input Format:
~~[INST] {prompt} [/INST]
LLM NameMistral 7B Instruct V0.1 GPTQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GPTQ 
Model NameMistral 7B Instruct v0.1
Model CreatorMistral AI
Base Model(s)  mistralai/Mistral-7B-Instruct-v0.1   mistralai/Mistral-7B-Instruct-v0.1
Model Size7b
Required VRAM4.2 GB
Updated2024-12-21
MaintainerTheBloke
Model Typemistral
Instruction-BasedYes
Model Files  4.2 GB
GPTQ QuantizationYes
Quantization Typegptq
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.34.0.dev0
Tokenizer ClassLlamaTokenizer
Vocabulary Size32000
Torch Data Typebfloat16

Quantized Models of the Mistral 7B Instruct V0.1 GPTQ

Model
Likes
Downloads
VRAM
Mistral Finetuned DialogSumm150 GB

Best Alternatives to Mistral 7B Instruct V0.1 GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Mistral 7B Instruct V0.2 GPTQ32K / 4.2 GB57367350
Mistral 7B Instruct V0.3 GPTQ32K / 4.2 GB155780
...ral 7B Instruct V0.3 GPTQ 4bit32K / 4.2 GB137116
Mistral 7B Instruct V0.3 GPTQ32K / 4.2 GB110
...tral 7B Instruct 32K GPTQ INT832K / 7.7 GB151
...ral 7B Instruct V0.3 GPTQ 4bit32K / 4.2 GB150
...stral 7B Instruct V0.3 GPTQ 8B32K / 7.7 GB130
...stral 7B Instruct V0.3 GPTQ 4B32K / 4.2 GB80
...nstruct V0.2 Neural Story GPTQ32K / 4.6 GB171
...olphin 2.6 Mistral 7B DPO GPTQ32K / 4.2 GB2269
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Mistral-7B-Instruct-v0.1-GPTQ.

Rank the Mistral 7B Instruct V0.1 GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217