Llama 2 7B GPTQ by localmodels

 Β»  All LLMs  Β»  localmodels  Β»  Llama 2 7B GPTQ   URL Share it on

  4bit   Autotrain compatible   Endpoints compatible   Gptq   Llama   Quantized   Region:us

Llama 2 7B GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 2 7B GPTQ (localmodels/Llama-2-7B-GPTQ)

Llama 2 7B GPTQ Parameters and Internals

Model Type 
auto-regressive, generative text model
Use Cases 
Areas:
Commercial and research applications in English
Applications:
Natural language generation tasks, Assistant-like chat
Primary Use Cases:
Pretrained models can be adapted for various tasks
Limitations:
Use in languages other than English, Violates laws or regulations
Considerations:
Specific formatting required to get expected features for chat.
Additional Notes 
Carbon footprint of pretraining is offset by Meta’s program.
Supported Languages 
languages_zero_shot (/0/), proficiency_level (/0/), default_language (/0/)
Training Details 
Data Sources:
A new mix of publicly available online data
Data Volume:
2 trillion tokens
Methodology:
Pretrained using auto-regressive architecture and fine-tuned with supervised learning and reinforcement learning with human feedback.
Context Length:
4096
Training Time:
January 2023 - July 2023
Hardware Used:
Meta's Research Super Cluster, third-party cloud compute
Model Architecture:
Optimized transformer architecture
Safety Evaluation 
Methodologies:
Evaluation on standard academic benchmarks
Findings:
Outperform open-source chat models on most benchmarks tested, Par with some closed-source models
Ethical Considerations:
Before deploying applications, developers should perform safety testing tailored to specific applications.
Responsible Ai Considerations 
Fairness:
Testing covers English scenarios, cannot predict nor cover all scenarios.
Accountability:
Developers should perform safety testing tailored to specific applications.
Mitigation Strategies:
Tuned with reinforcement learning with human feedback for alignment.
Input Output 
Input Format:
Text only
Accepted Modalities:
text
Output Format:
Text only
Performance Tips:
Bigger models (70B) use Grouped-Query Attention for improved scalability.
LLM NameLlama 2 7B GPTQ
Repository πŸ€—https://huggingface.co/localmodels/Llama-2-7B-GPTQ 
Base Model(s)  Llama 2 7B PetGPT   sudip2003/llama-2-7b-PetGPT
Model Size7b
Required VRAM3.9 GB
Updated2025-02-22
Maintainerlocalmodels
Model Typellama
Model Files  3.9 GB
GPTQ QuantizationYes
Quantization Typegptq|4bit
Model ArchitectureLlamaForCausalLM
Context Length2048
Model Max Length2048
Transformers Version4.30.2
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typefloat16

Best Alternatives to Llama 2 7B GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Yarn Llama 2 7B 128K GPTQ128K / 3.9 GB807
Yarn Llama 2 7B 64K GPTQ64K / 3.9 GB441
... 7B 32K Instructions V4 Marlin32K / 4.1 GB60
Aixcoder 7B GPTQ32K / 4.5 GB771
Calm2 7B Chat GPTQ32K / 4.4 GB716
...Calm2 7B Chat GPTQ Calib Ja 1K32K / 4.4 GB185
Llama 2 7B 32K Instruct GPTQ32K / 3.9 GB5727
Codebear 7B 4bit16K / 3.9 GB41
...a 7B Instruct GPTQ Calib Ja 1K16K / 3.9 GB120
CodeLlama 7B Instruct GPTQ16K / 3.9 GB59246
Note: green Score (e.g. "73.2") means that the model is better than localmodels/Llama-2-7B-GPTQ.

Rank the Llama 2 7B GPTQ Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227