Meta Llama 3 8B by NousResearch

 Β»  All LLMs  Β»  NousResearch  Β»  Meta Llama 3 8B   URL Share it on

  Autotrain compatible   En   Endpoints compatible   Facebook   Llama   Llama-3   Meta   Pytorch   Region:us   Safetensors   Sharded   Tensorflow

Meta Llama 3 8B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Meta Llama 3 8B (NousResearch/Meta-Llama-3-8B)

Meta Llama 3 8B Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
Commercial, Research
Applications:
Instruction tuned models for assistant-like chat
Primary Use Cases:
Natural language generation, Multilingual dialogue interactions
Limitations:
Out-of-the-box use only in English, Potential inaccurate or biased responses
Considerations:
Developers should fine-tune based on specific needs.
Additional Notes 
100% carbon emissions offset by Meta’s sustainability program.
Supported Languages 
en (high)
Training Details 
Data Sources:
publicly available online data
Data Volume:
15 trillion tokens
Methodology:
Supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF)
Context Length:
8000
Hardware Used:
H100-80GB GPU with a cumulative 7.7M GPU hours
Model Architecture:
Auto-regressive transformer architecture
Safety Evaluation 
Methodologies:
Red teaming exercises, Adversarial evaluations
Risk Categories:
CBRNE, Cyber Security, Child Safety
Ethical Considerations:
Leverages best practices for safety and responsible deployment.
Responsible Ai Considerations 
Fairness:
Inclusive and open approach, aiming to serve diverse user needs and perspectives.
Accountability:
Developers responsible for end-user safety evaluations.
Mitigation Strategies:
Tools like Meta Llama Guard 2 and Code Shield for layering safety measures.
Input Output 
Input Format:
text
Accepted Modalities:
text
Output Format:
text and code
Performance Tips:
Fine-tune with language-specific data where appropriate.
Release Notes 
Version:
Meta Llama 3 (8B, 70B)
Date:
April 18, 2024
Notes:
Initial release of pre-trained and instruction tuned variants.
LLM NameMeta Llama 3 8B
Repository πŸ€—https://huggingface.co/NousResearch/Meta-Llama-3-8B 
Model Size8b
Required VRAM16.1 GB
Updated2025-02-05
MaintainerNousResearch
Model Typellama
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length8192
Model Max Length8192
Transformers Version4.40.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size128256
Torch Data Typebfloat16

Quantized Models of the Meta Llama 3 8B

Model
Likes
Downloads
VRAM
...mes 2 Pro Llama 3 8B Bpw6 EXL2056 GB
Hermes 2 Pro Llama 3 8B Marlin145 GB
...ta Llama 3 8B HQQ 4bit Smashed0105 GB
...ta Llama 3 8B HQQ 2bit Smashed054 GB

Best Alternatives to Meta Llama 3 8B

Best Alternatives
Context / RAM
Downloads
Likes
...a 3 8B Instruct Gradient 1048K1024K / 16.1 GB6623678
MrRoboto ProLong 8B V4i1024K / 16.1 GB661
...o ProLongBASE Pt8 Unaligned 8B1024K / 16.1 GB240
Mpasila Viking 8B1024K / 16.1 GB590
41024K / 16.1 GB3220
Thor V1.4 8B DARK FICTION1024K / 16.1 GB9412
161024K / 16.1 GB1690
Because Im Bored Nsfw11024K / 16.1 GB661
111024K / 16.1 GB1130
NBeerbower Narrative 8B 64K1024K / 16.1 GB321
Note: green Score (e.g. "73.2") means that the model is better than NousResearch/Meta-Llama-3-8B.

Rank the Meta Llama 3 8B Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227