CodeLlama 7B Instruct AWQ by TheBloke

 Β»  All LLMs  Β»  TheBloke  Β»  CodeLlama 7B Instruct AWQ   URL Share it on

  Arxiv:2308.12950   4-bit   Autotrain compatible   Awq   Code   Codegen   Instruct   Llama   Llama2   Quantized   Region:us   Safetensors

CodeLlama 7B Instruct AWQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
CodeLlama 7B Instruct AWQ (TheBloke/CodeLlama-7B-Instruct-AWQ)

CodeLlama 7B Instruct AWQ Parameters and Internals

Model Type 
llama
Use Cases 
Primary Use Cases:
Code completion, Infilling, Instructions/chat
Limitations:
Use in languages other than English is out of scope
Considerations:
Code Llama’s potential outputs cannot be predicted in advance, and the model may produce inaccurate or objectionable responses.
Additional Notes 
Supports AWQ and GPTQ quantization methods for reduced computational footprint.
Training Details 
Data Sources:
Evol Instruct Code
Data Volume:
Code Llama and its variants have been trained between January 2023 and July 2023.
Methodology:
Optimized transformer architecture
Context Length:
4096
Training Time:
400K GPU hours
Hardware Used:
A100-80GB GPUs
Model Architecture:
Auto-regressive language model
Responsible Ai Considerations 
Mitigation Strategies:
Developers should perform safety testing and tuning tailored to their specific applications of the model.
Input Output 
Input Format:
Models input text only
Accepted Modalities:
text
Output Format:
Models generate text only
LLM NameCodeLlama 7B Instruct AWQ
Repository πŸ€—https://huggingface.co/TheBloke/CodeLlama-7B-Instruct-AWQ 
Model NameCodeLlama 7B Instruct
Model CreatorMeta
Base Model(s)  codellama/CodeLlama-7b-instruct-hf   codellama/CodeLlama-7b-instruct-hf
Model Size7b
Required VRAM3.9 GB
Updated2024-12-22
MaintainerTheBloke
Model Typellama
Instruction-BasedYes
Model Files  3.9 GB
Supported Languagescode
AWQ QuantizationYes
Quantization Typeawq
Generates CodeYes
Model ArchitectureLlamaForCausalLM
Licensellama2
Context Length16384
Model Max Length16384
Transformers Version4.33.0.dev0
Tokenizer ClassCodeLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32016
Torch Data Typebfloat16

Best Alternatives to CodeLlama 7B Instruct AWQ

Best Alternatives
Context / RAM
Downloads
Likes
...ama 7B Instruct Hf W4 G128 AWQ16K / 3.9 GB220
...ruct Solidity Bnb 4bit Smashed16K / 4.2 GB140
...B Instruct Hf Bnb 4bit Smashed16K / 4.2 GB210
CodelLama7B Inst DPO 7K Mlx16K / 4.2 GB82
...eLlama 7B Instruct Hf 4bit MLX16K / 4.2 GB121
...6.7B Instruct 8.0bpw H8 EXL2 216K / 6.8 GB92
...6.7B Instruct 3.0bpw H6 EXL2 216K / 2.8 GB91
... 7B Instruct Nf4 Fp16 Upscaled16K / 13.5 GB150
CodeLlama 7B Instruct Fp1616K / 13.5 GB338
...Llama 7B Instruct Bf16 Sharded16K / 13.5 GB161
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/CodeLlama-7B-Instruct-AWQ.

Rank the CodeLlama 7B Instruct AWQ Capabilities

πŸ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217