Taiwan LLaMa V1.0 4bits GPTQ by weiren119

 ยป  All LLMs  ยป  weiren119  ยป  Taiwan LLaMa V1.0 4bits GPTQ   URL Share it on

  Arxiv:2104.09864   Arxiv:2305.13711   4bit   Autotrain compatible Dataset:yentinglin/traditional...   Dataset:yentinglin/zh tw c4   Gptq   Instruct   Llama   Quantized   Region:us   Zh

Taiwan LLaMa V1.0 4bits GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Taiwan LLaMa V1.0 4bits GPTQ (weiren119/Taiwan-LLaMa-v1.0-4bits-GPTQ)

Taiwan LLaMa V1.0 4bits GPTQ Parameters and Internals

Model Type 
Instruction-Tuned, Causal Language Model
Use Cases 
Applications:
Taiwanese culture and related applications, Instruction-following conversations
Supported Languages 
zh (Traditional Chinese)
Training Details 
Data Sources:
yentinglin/zh_TW_c4, yentinglin/traditional_chinese_instructions
Data Volume:
5 billion tokens; 490k conversations
Methodology:
Fine-tuned on conversational data
Context Length:
4096
Model Architecture:
Transformer architecture, leveraging flash attention 2, and bfloat16
LLM NameTaiwan LLaMa V1.0 4bits GPTQ
Repository ๐Ÿค—https://huggingface.co/weiren119/Taiwan-LLaMa-v1.0-4bits-GPTQ 
Model NameLanguage Models for Taiwanese Culture 1.0
Model CreatorYen-Ting Lin
Required VRAM7.3 GB
Updated2024-12-22
Maintainerweiren119
Model Typellama
Instruction-BasedYes
Model Files  7.3 GB   7.3 GB
Supported Languageszh
GPTQ QuantizationYes
Quantization Typegptq|4bit
Model ArchitectureLlamaForCausalLM
Licensellama2
Context Length4096
Model Max Length4096
Transformers Version4.31.0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typefloat16

Best Alternatives to Taiwan LLaMa V1.0 4bits GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Taiwan LLaMa V1.0 4bits GPTQ4K / 7.3 GB121
Phi 3.5 Instruct Vul128K / 7.6 GB120
...ene Phi 3.5 Mini Instruct Orpo128K / 7.6 GB101
...Instruct DPO V2.8.0bpw H8 EXL24K / 10.9 GB92
Reverse Instruct32K / 27 GB353
Codellama Cairo Instruct GGUF16K / 4.1 GB371
Etri Ones Solar4K / 42.9 GB750
Law Chat4K / 27 GB121432
Phi 3 Orpo V8.164K / 7.6 GB210
Small Instruct4K / 2.9 GB48721
Note: green Score (e.g. "73.2") means that the model is better than weiren119/Taiwan-LLaMa-v1.0-4bits-GPTQ.

Rank the Taiwan LLaMa V1.0 4bits GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217