OLMo 7B SFT by allenai

 ยป  All LLMs  ยป  allenai  ยป  OLMo 7B SFT   URL Share it on

  Arxiv:2402.00838   Autotrain compatible   Conversational   Custom code   Dataset:allenai/dolma Dataset:allenai/tulu-v2-sft-mi...   En   Hf olmo   Pytorch   Region:us   Sharded
Model Card on HF ๐Ÿค—: https://huggingface.co/allenai/OLMo-7B-SFT 

OLMo 7B SFT Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
OLMo 7B SFT (allenai/OLMo-7B-SFT)

OLMo 7B SFT Parameters and Internals

Model Type 
Transformer style autoregressive language model
Additional Notes 
Installation requires `ai2-olmo`. Use with HuggingFace Transformers<=4.39 for compatibility.
Supported Languages 
en (Proficient)
Training Details 
Data Sources:
allenai/dolma, allenai/tulu-v2-sft-mixture, allenai/ultrafeedback_binarized_cleaned
Data Volume:
2.5 Trillion Tokens
Methodology:
SFT
Context Length:
2048
Model Architecture:
Based on OLMo core architecture.
Responsible Ai Considerations 
Accountability:
Users are responsible for the outputs produced by the model.
Mitigation Strategies:
Developers should consider implementing safeguards for biases, privacy, and other potential harms.
LLM NameOLMo 7B SFT
Repository ๐Ÿค—https://huggingface.co/allenai/OLMo-7B-SFT 
Model Size7b
Required VRAM13.7 GB
Updated2025-02-05
Maintainerallenai
Model Typehf_olmo
Model Files  9.9 GB: 1-of-2   3.8 GB: 2-of-2
Supported Languagesen
Model ArchitectureOLMoForCausalLM
Licenseapache-2.0
Transformers Version4.35.0.dev0
Tokenizer ClassOLMoTokenizer
Padding Token<|padding|>
Vocabulary Size50280
Torch Data Typebfloat16

Best Alternatives to OLMo 7B SFT

Best Alternatives
Context / RAM
Downloads
Likes
OLMo 7B0K / 27.6 GB25058629
OLMo 7B Instruct0K / 13.7 GB141851
OLMo 7B 04240K / 27.6 GB7447
OLMo 1.7 7B0K / 27.6 GB4643
OLMo 7B Twin 2T0K / 27.6 GB12922
Note: green Score (e.g. "73.2") means that the model is better than allenai/OLMo-7B-SFT.

Rank the OLMo 7B SFT Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227