Swallow 7B Instruct GPTQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Swallow 7B Instruct GPTQ   URL Share it on

  4-bit   Autotrain compatible Base model:quantized:tokyotech... Base model:tokyotech-llm/swall...   En   Gptq   Instruct   Ja   Llama   Quantized   Region:us   Safetensors

Swallow 7B Instruct GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Swallow 7B Instruct GPTQ (TheBloke/Swallow-7B-Instruct-GPTQ)

Swallow 7B Instruct GPTQ Parameters and Internals

Model Type 
llama, text-generation
Use Cases 
Areas:
research, commercial applications
Limitations:
The models are still in early stages and might not align well with human intent or safety considerations.
Additional Notes 
Trained primarily with additional Japanese language data to increase efficiency and representation.
Supported Languages 
Japanese (High), English (High)
Training Details 
Data Sources:
Japanese Wikipedia, RefinedWeb, Swallow Corpus, The Pile
Methodology:
Continual pre-training with supervised fine-tuning (SFT)
Context Length:
4096
Model Architecture:
llama2
Input Output 
Accepted Modalities:
text
Output Format:
text
LLM NameSwallow 7B Instruct GPTQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/Swallow-7B-Instruct-GPTQ 
Model NameSwallow 7B Instruct
Model Creatortokyotech-llm
Base Model(s)  Swallow 7B Instruct Hf   tokyotech-llm/Swallow-7b-instruct-hf
Model Size7b
Required VRAM4.1 GB
Updated2024-12-22
MaintainerTheBloke
Model Typellama
Instruction-BasedYes
Model Files  4.1 GB
Supported Languagesen ja
GPTQ QuantizationYes
Quantization Typegptq
Model ArchitectureLlamaForCausalLM
Licensellama2
Context Length4096
Model Max Length4096
Transformers Version4.35.2
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size43176
Torch Data Typebfloat16

Best Alternatives to Swallow 7B Instruct GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
... 7B 32K Instructions V4 Marlin32K / 4.1 GB120
Llama 2 7B 32K Instruct GPTQ32K / 3.9 GB7327
...a 7B Instruct GPTQ Calib Ja 1K16K / 3.9 GB360
CodeLlama 7B Instruct GPTQ16K / 3.9 GB50046
...echless Tora Code 7B V1.0 GPTQ16K / 3.9 GB221
CausalLM 7B GPTQ8K / 5.9 GB595
Leo Hessianai 7B Chat GPTQ8K / 3.9 GB530
...ssianai 7B Chat Bilingual GPTQ8K / 3.9 GB240
Xgen 7B 8K Open Instruct Gptq8K / 4.2 GB134
Deepseek Math 7B Instruct GPTQ4K / 4.8 GB170
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Swallow-7B-Instruct-GPTQ.

Rank the Swallow 7B Instruct GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40123 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217