OLMo 2 1124 13B Instruct by allenai

 Β»  All LLMs  Β»  allenai  Β»  OLMo 2 1124 13B Instruct   URL Share it on

  Arxiv:2411.15124   Autotrain compatible Base model:allenai/olmo-2-1124... Base model:finetune:allenai/ol...   Conversational Dataset:allenai/rlvr-gsm-math-...   En   Endpoints compatible   Instruct   Olmo2   Region:us   Safetensors   Sharded   Tensorflow

OLMo 2 1124 13B Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
OLMo 2 1124 13B Instruct (allenai/OLMo-2-1124-13B-Instruct)

OLMo 2 1124 13B Instruct Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
research, education
Additional Notes 
OLMo is a series of open language models.
Supported Languages 
en (primary)
Training Details 
Data Sources:
allenai/tulu-3-sft-olmo-2-mixture, allenai/olmo-2-1124-13b-preference-mix, allenai/RLVR-GSM-MATH-IF-Mixed-Constraints
Methodology:
supervised finetuning on TΓΌlu 3 dataset, DPO training, RLVR training
Safety Evaluation 
Risk Categories:
bias
Ethical Considerations:
Limited safety training, potential for problematic outputs.
Input Output 
Accepted Modalities:
text
Output Format:
text
Release Notes 
Date:
November 2024
Notes:
Post-trained variant with RLVR training.
LLM NameOLMo 2 1124 13B Instruct
Repository πŸ€—https://huggingface.co/allenai/OLMo-2-1124-13B-Instruct 
Base Model(s)  OLMo 2 1124 13B DPO   allenai/OLMo-2-1124-13B-DPO
Model Size13b
Required VRAM27.4 GB
Updated2024-12-08
Maintainerallenai
Model Typeolmo2
Instruction-BasedYes
Model Files  5.0 GB: 1-of-6   5.0 GB: 2-of-6   4.9 GB: 3-of-6   4.9 GB: 4-of-6   4.9 GB: 5-of-6   2.7 GB: 6-of-6
Supported Languagesen
Model ArchitectureOlmo2ForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.47.0.dev0
Tokenizer ClassGPT2Tokenizer
Padding Token<|pad|>
Vocabulary Size100352
Torch Data Typebfloat16
Errorsreplace

Rank the OLMo 2 1124 13B Instruct Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 39016 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124