Colossal LLaMA 2 7B Base by hpcai-tech

 ยป  All LLMs  ยป  hpcai-tech  ยป  Colossal LLaMA 2 7B Base   URL Share it on

  Arxiv:2110.14883   Arxiv:2307.09288   Autotrain compatible   En   Endpoints compatible   Llama   Pytorch   Region:us   Zh

Colossal LLaMA 2 7B Base Benchmarks

Colossal LLaMA 2 7B Base (hpcai-tech/Colossal-LLaMA-2-7b-base)

Colossal LLaMA 2 7B Base Parameters and Internals

Model Type 
language model, multilingual
Use Cases 
Areas:
Research, Commercial applications
Limitations:
May generate inaccurate or biased responses, Limited testing to English and Chinese use cases
Considerations:
Developers should undertake safety testing prior to deployment.
Additional Notes 
Designed for scalability and extensive language support with a tailored tokenizer for better semantic capture.
Supported Languages 
zh (high), en (high)
Training Details 
Data Volume:
8.5 billion tokens
Methodology:
Multi-stage training with bucket-based data placement
Context Length:
4096
Training Time:
15 hours
Hardware Used:
64 A800 GPUs
Input Output 
Input Format:
Text
Accepted Modalities:
text
Output Format:
Generated textual sequences
Performance Tips:
Utilize the model's expansive context window for improved coherence in long responses.
Release Notes 
Version:
2-7B-base
Notes:
Initial release based on LLaMA-2 model series with enhanced multilingual capabilities.
LLM NameColossal LLaMA 2 7B Base
Repository ๐Ÿค—https://huggingface.co/hpcai-tech/Colossal-LLaMA-2-7b-base 
Model Size7b
Required VRAM0.8 GB
Updated2025-06-01
Maintainerhpcai-tech
Model Typellama
Model Files  1.1 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   1.0 GB   0.8 GB
Supported Languageszh en
Model ArchitectureLlamaForCausalLM
Licensellama2
Context Length4096
Model Max Length4096
Transformers Version4.31.0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size69104
Torch Data Typefloat16

Best Alternatives to Colossal LLaMA 2 7B Base

Best Alternatives
Context / RAM
Downloads
Likes
A6 L1024K / 16.1 GB2010
M1024K / 16.1 GB1270
1571024K / 16.1 GB1010
1241024K / 16.1 GB930
A3.41024K / 16.1 GB130
A5.41024K / 16.1 GB120
A2.41024K / 16.1 GB120
2 Very Sci Fi1024K / 16.1 GB3170
1621024K / 16.1 GB600
1181024K / 16.1 GB150
Note: green Score (e.g. "73.2") means that the model is better than hpcai-tech/Colossal-LLaMA-2-7b-base.

Rank the Colossal LLaMA 2 7B Base Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 47770 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227