Llama 2 7B Chat GGUF by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Llama 2 7B Chat GGUF   URL Share it on

  Arxiv:2307.09288 Base model:meta-llama/llama-2-... Base model:quantized:meta-llam...   En   Facebook   Gguf   Llama   Llama2   Meta   Pytorch   Quantized   Region:us

Llama 2 7B Chat GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

Llama 2 7B Chat GGUF Parameters and Internals

Model Type 
text generation, dialogue
Use Cases 
Areas:
Commercial applications, Research
Primary Use Cases:
Assistant-like chat and dialogue
Limitations:
Unpredictable responses possible
Considerations:
Ensure safety testing and alignment to specific use.
Additional Notes 
Training and fine-tuning do not use Meta user data. Pretraining data cut off September 2022. Some tuning data up to July 2023.
Supported Languages 
en (high proficiency)
Training Details 
Data Sources:
A new mix of publicly available online data
Data Volume:
2.0 trillion tokens
Methodology:
Uses supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF)
Context Length:
4096
Hardware Used:
Meta's Research Super Cluster and production clusters for pretraining, A100-80GB GPUs for a total of 3.3M GPU hours
Model Architecture:
Auto-regressive language model with optimized transformer architecture
Safety Evaluation 
Methodologies:
internal evaluations library
Findings:
Llama 2 7B scores well on TruthfulQA and presents low toxicity on ToxiGen
Risk Categories:
misinformation, bias
Ethical Considerations:
Model output may be unpredictable. Safety testing and tuning recommended before deployment.
Responsible Ai Considerations 
Fairness:
Testing conducted in English. May not cover all scenarios.
Transparency:
Details available in Responsible Use Guide on website.
Accountability:
Reports via given platforms
Mitigation Strategies:
Internal evaluations and updates
Input Output 
Input Format:
Text input with specific formatting.
Accepted Modalities:
text
Output Format:
Text generation
Performance Tips:
Use INST and <> tags for expected performance.
LLM NameLlama 2 7B Chat GGUF
Repository ๐Ÿค—https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF 
Model NameLlama 2 7B Chat
Model CreatorMeta Llama 2
Base Model(s)  Llama 2 7B Chat Hf   meta-llama/Llama-2-7b-chat-hf
Model Size7b
Required VRAM2.8 GB
Updated2024-11-21
MaintainerTheBloke
Model Typellama
Model Files  2.8 GB   3.6 GB   3.3 GB   3.0 GB   3.8 GB   4.1 GB   3.9 GB   4.7 GB   4.8 GB   4.7 GB   5.5 GB   7.2 GB
Supported Languagesen
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureAutoModel
Licensellama2
Llama 2 7B Chat GGUF (TheBloke/Llama-2-7B-Chat-GGUF)

Best Alternatives to Llama 2 7B Chat GGUF

Best Alternatives
Context / RAM
Downloads
Likes
Pixel8K / 4.4 GB180
Mistral 7B Instruct V0.3 GGUF0K / 1.6 GB197055363
Qwen2 7B Instruct GGUF0K / 1.9 GB21305429
WizardLM 2 7B GGUF0K / 2.7 GB190603474
Mistral 7B Instruct V0.2 GGUF0K / 3.1 GB138341398
Mistral 7B Instruct V0.3 GGUF0K / 2.7 GB583277
Qwen2 7B Instruct V0.6 GGUF0K / 4.5 GB135220
Mistral 7B Instruct V0.1 GGUF0K / 3.1 GB188922511
Qwen2 7B Instruct V0.1 GGUF0K / 4.5 GB97140
Qwen2 7B Instruct V0.7 GGUF0K / 4.5 GB95300
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Llama-2-7B-Chat-GGUF.

Rank the Llama 2 7B Chat GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 38149 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241110