Galactica 125M by facebook

 ยป  All LLMs  ยป  facebook  ยป  Galactica 125M   URL Share it on

  Arxiv:1810.03993   Autotrain compatible   Galactica   Opt   Pytorch   Region:us   Safetensors
Model Card on HF ๐Ÿค—: https://huggingface.co/facebook/galactica-125m 

Galactica 125M Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Galactica 125M (facebook/galactica-125m)

Galactica 125M Parameters and Internals

Model Type 
Transformer based architecture, Decoder-only
Use Cases 
Areas:
research, scientific tooling development
Primary Use Cases:
citation prediction, scientific QA, mathematical reasoning, summarization, document generation, molecular property prediction, entity extraction
Limitations:
prone to hallucination, popularity bias in citation behavior
Training Details 
Data Sources:
papers, textbooks, scientific websites, encyclopedias, reference material, knowledge bases
Data Volume:
106 billion tokens
LLM NameGalactica 125M
Repository ๐Ÿค—https://huggingface.co/facebook/galactica-125m 
Model Size125m
Required VRAM0.2 GB
Updated2025-03-18
Maintainerfacebook
Model Typeopt
Model Files  0.2 GB   0.2 GB
Model ArchitectureOPTForCausalLM
Licensecc-by-nc-4.0
Context Length2048
Model Max Length2048
Transformers Version4.21.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size50000
Torch Data Typefloat32
Activation Functiongelu

Best Alternatives to Galactica 125M

Best Alternatives
Context / RAM
Downloads
Likes
...125M Qcqa Ub 6 Best For Q Loss2K / 0.5 GB17610
...5M Qcqa Ub 6 Best For KV Cache2K / 0.5 GB17570
...25M Gqa Ub 6 Best For KV Cache2K / 0.5 GB17750
Opt 125M2K / 0.3 GB9092506187
Galactica 125M Cot2K / 0.5 GB520
Galactica Ref2K / 0.5 GB60
Galactica 125M DPO Pos2K / 0.5 GB150
Galactica 125M DPO2K / 0.5 GB140
BertQA2K / 0.5 GB260
BertQA2K / 0.5 GB270
Note: green Score (e.g. "73.2") means that the model is better than facebook/galactica-125m.

Rank the Galactica 125M Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 45269 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227