Minitron 4B Base by nvidia

 ยป  All LLMs  ยป  nvidia  ยป  Minitron 4B Base   URL Share it on

  Arxiv:2009.03300   Arxiv:2407.14679   Nemo   Nemotron   Pytorch   Region:us
Model Card on HF ๐Ÿค—: https://huggingface.co/nvidia/Minitron-4B-Base 

Minitron 4B Base Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Minitron 4B Base (nvidia/Minitron-4B-Base)

Minitron 4B Base Parameters and Internals

Model Type 
Transformer Decoder, auto-regressive language model
Use Cases 
Areas:
research, development
Limitations:
amplifies biases, can generate toxic responses, may produce inaccurate or undesirable text
Considerations:
Work with internal model team to ensure model meets industry requirements and addresses unforeseen misuse.
Supported Languages 
en (basic proficiency), multilingual (basic proficiency)
Training Details 
Data Sources:
webpages, dialogue, articles, other written materials
Data Volume:
94 billion tokens
Methodology:
pruning and knowledge distillation
Model Architecture:
Transformer Decoder
Input Output 
Input Format:
String
Accepted Modalities:
Text
Output Format:
String
Performance Tips:
Load the model using TensorRT-LLM on supported hardware with CUDA and Torch settings for optimal performance.
LLM NameMinitron 4B Base
Repository ๐Ÿค—https://huggingface.co/nvidia/Minitron-4B-Base 
Model Size4b
Required VRAM8.4 GB
Updated2024-12-22
Maintainernvidia
Model Typenemotron
Model Files  8.4 GB
Model ArchitectureNemotronForCausalLM
Licenseother
Context Length4096
Model Max Length4096
Transformers Version4.32.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size256000
Torch Data Typebfloat16

Best Alternatives to Minitron 4B Base

Best Alternatives
Context / RAM
Downloads
Likes
Nemotron Mini 4B Instruct4K / 8.4 GB45134
Note: green Score (e.g. "73.2") means that the model is better than nvidia/Minitron-4B-Base.

Rank the Minitron 4B Base Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217