Yi Ko 34B by beomi

 ยป  All LLMs  ยป  beomi  ยป  Yi Ko 34B   URL Share it on

  01-ai   Autotrain compatible   En   Ko   Llama   Pytorch   Region:us   Safetensors   Sharded   Tensorflow   Yi   Yi-ko
Model Card on HF ๐Ÿค—: https://huggingface.co/beomi/Yi-Ko-34B 

Yi Ko 34B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Yi Ko 34B (beomi/Yi-Ko-34B)

Yi Ko 34B Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
research applications
Applications:
text generation
Limitations:
text only input and output
Supported Languages 
Korean (advanced), English (advanced)
Training Details 
Data Sources:
A mix of Korean + English online data
Data Volume:
40B+ tokens
Context Length:
4000
Model Architecture:
Yi-Ko series models are an auto-regressive language model that uses an optimized transformer architecture based on Llama-2.
Input Output 
Input Format:
text
Accepted Modalities:
text
Output Format:
text
Release Notes 
Version:
2024.07.08
Date:
2024.07.08
Notes:
Update LICENSE to Apache 2.0
LLM NameYi Ko 34B
Repository ๐Ÿค—https://huggingface.co/beomi/Yi-Ko-34B 
Model Size34b
Required VRAM69.5 GB
Updated2025-02-22
Maintainerbeomi
Model Typellama
Model Files  2.8 GB: 1-of-25   3.0 GB: 2-of-25   2.9 GB: 3-of-25   2.8 GB: 4-of-25   2.8 GB: 5-of-25   2.8 GB: 6-of-25   2.8 GB: 7-of-25   2.8 GB: 8-of-25   2.8 GB: 9-of-25   2.8 GB: 10-of-25   2.8 GB: 11-of-25   2.8 GB: 12-of-25   2.8 GB: 13-of-25   2.8 GB: 14-of-25   2.8 GB: 15-of-25   2.8 GB: 16-of-25   2.8 GB: 17-of-25   2.8 GB: 18-of-25   2.8 GB: 19-of-25   2.8 GB: 20-of-25   2.8 GB: 21-of-25   2.8 GB: 22-of-25   2.8 GB: 23-of-25   2.8 GB: 24-of-25   2.0 GB: 25-of-25
Supported Languagesen ko
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.33.1
Tokenizer ClassLlamaTokenizer
Padding Token<unk>
Vocabulary Size78464
Torch Data Typebfloat16

Best Alternatives to Yi Ko 34B

Best Alternatives
Context / RAM
Downloads
Likes
Casual Magnum 34B195K / 68.8 GB141
34B Beta195K / 69.2 GB372963
Bagel 34B V0.2195K / 68.7 GB679040
Bagel Hermes 34B Slerp195K / 68.9 GB38941
Smaug 34B V0.1195K / 69.2 GB367260
Yi 34B 200K195K / 68.9 GB6015318
Yi 34B 200K AEZAKMI V2195K / 69.2 GB200712
Faro Yi 34B195K / 69.2 GB36126
Smaug 34B V0.1 ExPO195K / 69.2 GB19720
Mergekit Slerp Anaazls195K / 69.2 GB90
Note: green Score (e.g. "73.2") means that the model is better than beomi/Yi-Ko-34B.

Rank the Yi Ko 34B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227