Llama 3 8B GPTQ 4 Bit by astronomer

 ยป  All LLMs  ยป  astronomer  ยป  Llama 3 8B GPTQ 4 Bit   URL Share it on

  Arxiv:2210.17323   4-bit   Astronomer   Autotrain compatible Base model:meta-llama/meta-lla... Base model:quantized:meta-llam...   Dataset:wikitext   Endpoints compatible   Facebook   Finetuned   Gptq   Llama   Llama-3   Meta   Pretrained   Quantized   Region:us   Safetensors

Llama 3 8B GPTQ 4 Bit Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 3 8B GPTQ 4 Bit (astronomer/Llama-3-8B-GPTQ-4-Bit)

Llama 3 8B GPTQ 4 Bit Parameters and Internals

Model Type 
llama, text-generation
Additional Notes 
This model is quantized to 4-bit using GPTQ methodology by David Xue from Astronomer. It uses the wikitext dataset for calibration.
LLM NameLlama 3 8B GPTQ 4 Bit
Repository ๐Ÿค—https://huggingface.co/astronomer/Llama-3-8B-GPTQ-4-Bit 
Model NameMeta-Llama-3-8B
Model Creatorastronomer-io
Base Model(s)  Meta Llama 3 8B   meta-llama/Meta-Llama-3-8B
Model Size8b
Required VRAM5.7 GB
Updated2025-04-28
Maintainerastronomer
Model Typellama
Model Files  5.7 GB
GPTQ QuantizationYes
Quantization Typegptq
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length8192
Model Max Length8192
Transformers Version4.38.2
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size128256
Torch Data Typebfloat16

Best Alternatives to Llama 3 8B GPTQ 4 Bit

Best Alternatives
Context / RAM
Downloads
Likes
... 8B Instruct 262K 4bit GPTQ 02256K / 5.7 GB60
...a 3 8B Instruct 262K 4bit GPTQ256K / 5.8 GB21
...lama 3.1 8B Instruct GPTQ INT4128K / 5.8 GB12855225
...Instruct 80K Qlora Merged GPTQ80K / 5.8 GB50
...oLeo Instruct 8B 32K V0.1 GPTQ64K / 5.7 GB60
Llama3 German 8B 32K GPTQ64K / 5.7 GB50
Tsukasa Llama 3 8B Qlora Gptq32K / 5.8 GB50
Llama 3 Soliloquy 8B V2 GPTQ24K / 5.7 GB101
Llama 3 Soliloquy 8B GPTQ16K / 5.7 GB105
Meta Llama 3 8B Instruct GPTQ8K / 5.7 GB120563
Note: green Score (e.g. "73.2") means that the model is better than astronomer/Llama-3-8B-GPTQ-4-Bit.

Rank the Llama 3 8B GPTQ 4 Bit Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 46763 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227