SmolLM2 1.7B Instruct by HuggingFaceTB

 ยป  All LLMs  ยป  HuggingFaceTB  ยป  SmolLM2 1.7B Instruct   URL Share it on

  Autotrain compatible Base model:huggingfacetb/smoll... Base model:quantized:huggingfa...   Conversational   En   Endpoints compatible   Ext 8k   Instruct   Llama   Onnx   Region:us   Safetensors   Tensorboard   Transformers.js

SmolLM2 1.7B Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
SmolLM2 1.7B Instruct (HuggingFaceTB/SmolLM2-1.7B-Instruct)

SmolLM2 1.7B Instruct Parameters and Internals

Model Type 
text generation, instruction following
Use Cases 
Areas:
text generation, instruction following, text rewriting, summarization, function calling
Applications:
educational tools, coding assistance, customer support, chatbots, language learning
Primary Use Cases:
text and instruction generation
Limitations:
primarily understands and generates content in English, generated content may not always be factually accurate or free from bias
Considerations:
Users should always verify important information and critically evaluate any generated content.
Supported Languages 
en (main)
Training Details 
Data Sources:
FineWeb-Edu, DCLM, The Stack, new mathematics and coding datasets
Data Volume:
11T tokens
Methodology:
Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO)
Hardware Used:
256 H100 GPUs
Model Architecture:
Transformer decoder
Input Output 
Input Format:
Expected input format includes system and user prompts
Accepted Modalities:
text
Output Format:
text
Performance Tips:
Ensure queries are clear and check model's language support for optimal results.
LLM NameSmolLM2 1.7B Instruct
Repository ๐Ÿค—https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct 
Base Model(s)  HuggingFaceTB/SmolLM2-1.7B   HuggingFaceTB/SmolLM2-1.7B
Model Size1.7b
Required VRAM3.4 GB
Updated2024-12-22
MaintainerHuggingFaceTB
Model Typellama
Instruction-BasedYes
Model Files  3.4 GB   0.0 GB
Supported Languagesen
Context Length8k
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length8192
Model Max Length8192
Transformers Version4.42.3
Tokenizer ClassGPT2Tokenizer
Padding Token<|im_end|>
Vocabulary Size49152
Torch Data Typebfloat16

Quantized Models of the SmolLM2 1.7B Instruct

Model
Likes
Downloads
VRAM
SmolLM2 1.7B Instruct Bnb 4bit014911 GB

Best Alternatives to SmolLM2 1.7B Instruct

Best Alternatives
Context / RAM
Downloads
Likes
SmolTulu 1.7B Reinforced8K / 3.4 GB355
SmolLM2 Math IIO 1.7B Instruct8K / 3.4 GB1058
SmolLM2 1.7 Persona8K / 3.5 GB180
SmolLM2 1.7B Instruct8K / 3.4 GB9763
SmolLM2 1.7B Instruct8K / 3.4 GB1383
SmolLM 1.7B Instruct2K / 3.4 GB42724107
SmolLM 1.7B Instruct2K / 3.4 GB14810
SmolLM 1.7B2K / 3.4 GB8780
SmolLM 1.7B Instruct IFEval2K / 3.4 GB260
SmolLM2 1.7B Instruct Bnb 4bit8K / 1 GB14910

Rank the SmolLM2 1.7B Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217