Baichuan 13B Instruction by AlpachinoNLP

 ยป  All LLMs  ยป  AlpachinoNLP  ยป  Baichuan 13B Instruction   URL Share it on

  Arxiv:2104.09864   Arxiv:2108.12409   Arxiv:2307.00360   Autotrain compatible   Baichuan   Custom code   En   Instruct   Pytorch   Region:us   Sharded   Zh

Baichuan 13B Instruction Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Baichuan 13B Instruction (AlpachinoNLP/Baichuan-13B-Instruction)

Baichuan 13B Instruction Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
Research, Text generation, Language tasks
Limitations:
Performance dependent on fine-tuning quality
Considerations:
Guidelines not specified
Additional Notes 
Baichuan-13B-Instruction includes ALiBi linear bias technology for improved inference speed metrics.
Supported Languages 
zh (Native), en (Intermediate)
Training Details 
Data Sources:
sharegpt_zh, lima, high-quality Chinese dataset selected by task type
Data Volume:
Not specified
Methodology:
Instruction fine-tuning involving specific high-quality datasets
Context Length:
4096
Training Time:
Not specified
Hardware Used:
8*A40 GPUs
Model Architecture:
Based on Baichuan-13B with ALiBi linear bias technology
Input Output 
Input Format:
Text prompts
Accepted Modalities:
text
Output Format:
Generated text
Performance Tips:
Quantization to int8 or int4 recommended for performance optimization.
LLM NameBaichuan 13B Instruction
Repository ๐Ÿค—https://huggingface.co/AlpachinoNLP/Baichuan-13B-Instruction 
Model Size13b
Required VRAM26.5 GB
Updated2025-02-05
MaintainerAlpachinoNLP
Model Typebaichuan
Instruction-BasedYes
Model Files  4.0 GB: 1-of-7   4.0 GB: 2-of-7   3.9 GB: 3-of-7   3.9 GB: 4-of-7   4.0 GB: 5-of-7   3.9 GB: 6-of-7   2.8 GB: 7-of-7
Supported Languageszh en
Model ArchitectureBaichuanForCausalLM
Model Max Length4096
Transformers Version4.29.2
Tokenizer ClassBaichuanTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size64000
Torch Data Typefloat16

Quantized Models of the Baichuan 13B Instruction

Model
Likes
Downloads
VRAM
Baichuan 13B Instruction GPTQ4167 GB

Best Alternatives to Baichuan 13B Instruction

Best Alternatives
Context / RAM
Downloads
Likes
... Efficient Training Of LLMs V10K / 29.1 GB341
NSFW 13B Sft0K / 26.5 GB10790
Blossom V2 Baichuan 13B0K / 26.5 GB161
Baichuan 13B Instruction GPTQ0K / 7.9 GB164
Note: green Score (e.g. "73.2") means that the model is better than AlpachinoNLP/Baichuan-13B-Instruction.

Rank the Baichuan 13B Instruction Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227