Sea Lion 7B by aisingapore

 ยป  All LLMs  ยป  aisingapore  ยป  Sea Lion 7B   URL Share it on

  Arxiv:2101.09635   Autotrain compatible   Custom code   En   Endpoints compatible   Fil   Id   Km   Lo   Mpt   Ms   My   Region:us   Safetensors   Sharded   Ta   Tensorflow   Th   Vi   Zh
Model Card on HF ๐Ÿค—: https://huggingface.co/aisingapore/sea-lion-7b 

Sea Lion 7B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Sea Lion 7B (aisingapore/sea-lion-7b)

Sea Lion 7B Parameters and Internals

Model Type 
Decoder
Additional Notes 
This is the repository for the base model. The model has not been aligned for safety.
Supported Languages 
en (English), zh (Chinese), id (Indonesian), ms (Malay), th (Thai), vi (Vietnamese), fil (Filipino), ta (Tamil), my (Burmese), km (Khmer), lo (Lao)
Training Details 
Data Sources:
RefinedWeb - English, mC4 - Chinese, mC4 - Indonesian, mC4 - Malay, mC4 - Filipino, mC4 - Burmese, mC4 - Vietnamese, mC4 - Thai, WangChanBERTa - Thai, mC4 - Lao, mC4 - Khmer, mC4 - Tamil, the Stack - Python, the Stack - Javascript, the Stack - Shell, the Stack - SQL, the Stack - Markdown, RedPajama - StackExchange, RedPajama - ArXiv
Data Volume:
980B tokens
Context Length:
2048
Training Time:
22 days
Hardware Used:
AWS EC2 p4d.24xlarge - 32 instances, Nvidia A100 40GB GPU - 256
Model Architecture:
Uses MPT architecture with 32 layers, d_model: 4096, head_dim: 32, Vocabulary: 256000, Sequence Length: 2048
LLM NameSea Lion 7B
Repository ๐Ÿค—https://huggingface.co/aisingapore/sea-lion-7b 
Model Size7b
Required VRAM15 GB
Updated2024-12-21
Maintaineraisingapore
Model Typempt
Model Files  9.9 GB: 1-of-2   5.1 GB: 2-of-2
Supported Languagesen zh id ms th vi ta my km lo
Model ArchitectureMPTForCausalLM
Licensemit
Transformers Version4.34.1
Tokenizer ClassSEABPETokenizer
Vocabulary Size256000
Torch Data Typebfloat16

Best Alternatives to Sea Lion 7B

Best Alternatives
Context / RAM
Downloads
Likes
Mpt 7B0K / 13.3 GB311371164
Mpt 7B Chat0K / 13.3 GB18663512
Mpt 7B Storywriter0K / 13.3 GB1817824
Mpt 7B Instruct0K / 13.3 GB8063468
Mpt 7B Int8 Ov0K / 0 GB100
Shears Mpt 7B 50 Base0K / 13.3 GB171
Sea Lion 7B Instruct0K / 15 GB53123
Mpt 7B0K / 26.5 GB36691
Mpt 7B 8K0K / 13.3 GB160426
Mpt 7B 8K Instruct0K / 13.3 GB132126
Note: green Score (e.g. "73.2") means that the model is better than aisingapore/sea-lion-7b.

Rank the Sea Lion 7B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40013 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217