Llama 3.1 Minitron 4B Width Base by nvidia

 ยป  All LLMs  ยป  nvidia  ยป  Llama 3.1 Minitron 4B Width Base   URL Share it on

  Arxiv:2009.03300   Arxiv:2407.14679   Arxiv:2408.11796   Autotrain compatible   En   Endpoints compatible   Llama   Llama-3   Nemo   Nvidia   Pytorch   Region:us   Safetensors   Sharded   Tensorflow

Llama 3.1 Minitron 4B Width Base Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 3.1 Minitron 4B Width Base (nvidia/Llama-3.1-Minitron-4B-Width-Base)

Llama 3.1 Minitron 4B Width Base Parameters and Internals

Model Type 
text-to-text, generative
Use Cases 
Areas:
research, commercial applications
Applications:
text generation, question answering
Primary Use Cases:
Paragraph completion
Limitations:
Potential for generating toxic, biased responses
Considerations:
Security measures and ethical reviews may be needed
Supported Languages 
English (high), Multilingual (medium)
Training Details 
Data Sources:
webpages, dialogue, articles
Data Volume:
94 billion tokens
Methodology:
Knowledge distillation
Context Length:
8000
Training Time:
July 29, 2024 and Aug 3, 2024
Hardware Used:
NVIDIA A100
Model Architecture:
Llama-3.1 core, Transformer Decoder
Responsible Ai Considerations 
Transparency:
Some known biases
Accountability:
User responsibility
Mitigation Strategies:
Ensure requirements for use cases are met, unexpected misuse addressed
Input Output 
Input Format:
String
Accepted Modalities:
Text
Output Format:
String
Performance Tips:
Use text prompts of 8000 characters or less for best results
Release Notes 
Version:
4B
Date:
2024-08-03
Notes:
Final deployment-ready version.
LLM NameLlama 3.1 Minitron 4B Width Base
Repository ๐Ÿค—https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base 
Model Size4b
Required VRAM9 GB
Updated2025-02-18
Maintainernvidia
Model Typellama
Model Files  5.0 GB: 1-of-2   4.0 GB: 2-of-2
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length131072
Model Max Length131072
Transformers Version4.45.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size128256
Torch Data Typebfloat16

Best Alternatives to Llama 3.1 Minitron 4B Width Base

Best Alternatives
Context / RAM
Downloads
Likes
SJT 4B146K / 7.6 GB160
Loxa 4B128K / 16 GB630
Aura 4B128K / 9 GB2610
...ama 3.1 Minitron 4B Depth Base128K / 9.1 GB502720
Nemotron W 4b MagLight 0.1128K / 9.2 GB101
....5 MINI 4B SFTxORPO HESSIAN AI128K / 7.7 GB160
....5 MINI 4B ORPOxSFT HESSIAN AI128K / 7.7 GB150
....5 MINI 4B ORPOxSFT HESSIAN AI128K / 7.7 GB150
....5 MINI 4B SFTxORPO HESSIAN AI128K / 7.7 GB130
DigitalSoul 4B128K / 9.3 GB223
Note: green Score (e.g. "73.2") means that the model is better than nvidia/Llama-3.1-Minitron-4B-Width-Base.

Rank the Llama 3.1 Minitron 4B Width Base Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43267 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227