Danube Ko 1.8B Base by jjhsnail0822

 ยป  All LLMs  ยป  jjhsnail0822  ยป  Danube Ko 1.8B Base   URL Share it on

  Arxiv:2402.14714   Autotrain compatible Base model:finetune:h2oai/h2o-... Base model:h2oai/h2o-danube2-1...   Dataset:uonlp/culturax   En   Endpoints compatible   H2o-danube2   Ko   Korean   Mistral   Region:us   Safetensors   Sllm

Danube Ko 1.8B Base Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Danube Ko 1.8B Base (jjhsnail0822/danube-ko-1.8b-base)

Danube Ko 1.8B Base Parameters and Internals

Model Type 
language model, pretrained model
Use Cases 
Considerations:
The Model can generate biased or inaccurate information. Use with caution.
Supported Languages 
ko (Korean), en (English)
Training Details 
Data Sources:
CulturaX, Common Crawl CC-MAIN-2024-10, AI Hub Data, Korean Wikis, Corpora from National Institute of the Korean Language, Standard Korean Dictionary
Data Volume:
42GB
Methodology:
EEVE technique
Context Length:
2048
Model Architecture:
Vocabulary size expanded from 32000 to 40000 for Korean tokens. Sequence length of 2048
LLM NameDanube Ko 1.8B Base
Repository ๐Ÿค—https://huggingface.co/jjhsnail0822/danube-ko-1.8b-base 
Base Model(s)  h2oai/h2o-danube2-1.8b-base   h2oai/h2o-danube2-1.8b-base
Model Size1.8b
Required VRAM3.7 GB
Updated2025-02-22
Maintainerjjhsnail0822
Model Typemistral
Model Files  3.7 GB
Supported Languagesko en
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.41.2
Tokenizer ClassLlamaTokenizer
Padding Token</s>
Vocabulary Size40000
Torch Data Typebfloat16

Best Alternatives to Danube Ko 1.8B Base

Best Alternatives
Context / RAM
Downloads
Likes
H2o Danube 1.8B Base16K / 3.7 GB39643
H2o Danube 1.8B Chat16K / 3.7 GB49754
Cypher Mini 1.8B16K / 3.7 GB1672
H2o Danube 1.8B Sft16K / 3.7 GB17311
Cypher CoT 1.8B16K / 3.7 GB1521
PixieZehirNano16K / 3.7 GB100
...1.8B Chat Sft Merge Fourier V116K / 7.3 GB901
H2o Danube2 1.8B Chat8K / 3.7 GB285661
H2o Danube2 1.8B Base8K / 3.7 GB24246
H2o Danube2 1.8B Sft8K / 3.7 GB2636
Note: green Score (e.g. "73.2") means that the model is better than jjhsnail0822/danube-ko-1.8b-base.

Rank the Danube Ko 1.8B Base Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227