OLMo 7B by allenai

 ยป  All LLMs  ยป  allenai  ยป  OLMo 7B   URL Share it on

  Arxiv:2302.13971   Arxiv:2402.00838   Autotrain compatible   Custom code   Dataset:allenai/dolma   En   Hf olmo   Pytorch   Region:us   Safetensors
Model Card on HF ๐Ÿค—: https://huggingface.co/allenai/OLMo-7B 

OLMo 7B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
OLMo 7B (allenai/OLMo-7B)

OLMo 7B Parameters and Internals

Model Type 
Transformer style autoregressive language model
Use Cases 
Areas:
research, commercial applications
Limitations:
easy to prompt models to generate harmful content, bias, many facts may not be true
Additional Notes 
OLMo models are available with open training and evaluation code.
Supported Languages 
English (NLP)
Training Details 
Data Sources:
allenai/dolma
Data Volume:
2.5 Trillion Tokens for OLMo 7B
Context Length:
2048
Hardware Used:
MI250X GPUs at the LUMI supercomputer, A100-40GB GPUs provided by MosaicML
Model Architecture:
OLMo 7B architecture with peer models for comparison: d_model 4096, num heads 32, num layers 32, activation SwiGLU
Input Output 
Input Format:
Text
Accepted Modalities:
text
Output Format:
Text generation
Release Notes 
Version:
1.0
Notes:
Core OLMo 7B release with model details, performance, and usage guidelines.
LLM NameOLMo 7B
Repository ๐Ÿค—https://huggingface.co/allenai/OLMo-7B 
Model Size7b
Required VRAM27.6 GB
Updated2025-02-05
Maintainerallenai
Model Typehf_olmo
Model Files  27.6 GB   27.6 GB
Supported Languagesen
Model ArchitectureOLMoForCausalLM
Licenseapache-2.0
Transformers Version4.36.2
Tokenizer ClassOLMoTokenizer
Padding Token<|padding|>
Vocabulary Size50280

Best Alternatives to OLMo 7B

Best Alternatives
Context / RAM
Downloads
Likes
OLMo 7B Instruct0K / 13.7 GB141851
OLMo 7B 04240K / 27.6 GB7447
OLMo 1.7 7B0K / 27.6 GB4643
OLMo 7B SFT0K / 13.7 GB763
OLMo 7B Twin 2T0K / 27.6 GB12922

Rank the OLMo 7B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227