CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2 by LoneStriker

 Β»  All LLMs  Β»  LoneStriker  Β»  CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2   URL Share it on

  Arxiv:2308.12950   Autotrain compatible   Code   Codegen   Endpoints compatible   Exl2   Instruct   Llama   Llama2   Pytorch   Quantized   Region:us   Safetensors   Sharded   Tensorflow

CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2 (LoneStriker/CodeLlama-70b-Instruct-hf-4.0bpw-h6-exl2)

CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2 Parameters and Internals

Model Type 
text generation, code synthesis
Use Cases 
Areas:
commercial use, research
Applications:
code synthesis, code understanding, Python code generation
Primary Use Cases:
instruction following, safer deployment in code generation
Limitations:
English only, Requires careful tuning for safety, Not suitable for legal or regulation-violating activities
Considerations:
Use in a way that adheres to the Responsible Use Guide.
Additional Notes 
Variation in model capabilities based on size and training.
Supported Languages 
English (proficient), Python (specialized)
Training Details 
Data Sources:
Offline datasets
Data Volume:
Large
Methodology:
Fine-tuning on instruct data
Context Length:
16000
Training Time:
Extended
Hardware Used:
Meta’s Research Super Cluster
Model Architecture:
Optimized transformer architecture
Safety Evaluation 
Methodologies:
safety evaluations outlined in the research paper
Findings:
Potential to produce inaccurate or objectionable responses.
Risk Categories:
misinformation, bias
Ethical Considerations:
Developers should perform safety testing and tuning tailored to their specific applications of the model.
Responsible Ai Considerations 
Fairness:
Testing has been primarily in English and cannot cover all scenarios.
Transparency:
Outputs cannot be predicted in advance, responsible use guide provided.
Accountability:
Developers should ensure applications comply with relevant use cases.
Mitigation Strategies:
Developers should perform safety testing tailored to their specific applications of the model.
Input Output 
Input Format:
text
Accepted Modalities:
text
Output Format:
text
Performance Tips:
Follow updated prompt template for 70B Instruct model.
LLM NameCodeLlama 70B Instruct Hf 4.0bpw H6 EXL2
Repository πŸ€—https://huggingface.co/LoneStriker/CodeLlama-70b-Instruct-hf-4.0bpw-h6-exl2 
Model Size70b
Required VRAM35.1 GB
Updated2024-12-22
MaintainerLoneStriker
Model Typellama
Instruction-BasedYes
Model Files  8.5 GB: 1-of-5   8.6 GB: 2-of-5   8.6 GB: 3-of-5   8.5 GB: 4-of-5   0.9 GB: 5-of-5
Supported Languagescode
Quantization Typeexl2
Generates CodeYes
Model ArchitectureLlamaForCausalLM
Licensellama2
Context Length2048
Model Max Length2048
Transformers Version4.37.1
Tokenizer ClassLlamaTokenizer
Vocabulary Size32016
Torch Data Typebfloat16

Best Alternatives to CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2

Best Alternatives
Context / RAM
Downloads
Likes
...70B Instruct Nf4 Fp16 Upscaled4K / 138.7 GB4092
...Llama 70B Instruct Hf 4bit MLX4K / 39.1 GB2425
...70B Instruct Hf 5.0bpw H6 EXL22K / 43.6 GB76
...0B Instruct Hf 2.65bpw H6 EXL22K / 23.4 GB93
...70B Instruct Hf 2.4bpw H6 EXL22K / 21.3 GB101
CodeLlama 70B Instruct Hf4K / 72.3 GB4291204
Code Llama 70B Python Instruct4K / 138.1 GB821
CodeLlama 70B Instruct Hf4K / 72.3 GB28616
CodeLlama 70B Instruct Neuron4K /  GB4251
CodeLlama 70B Instruct Hf GGUF4K / 25.5 GB1742
Note: green Score (e.g. "73.2") means that the model is better than LoneStriker/CodeLlama-70b-Instruct-hf-4.0bpw-h6-exl2.

Rank the CodeLlama 70B Instruct Hf 4.0bpw H6 EXL2 Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217