Qwen1.5 110B Chat GGUF by second-state

 ยป  All LLMs  ยป  second-state  ยป  Qwen1.5 110B Chat GGUF   URL Share it on

  Autotrain compatible Base model:qwen/qwen1.5-110b-c...   Chat   En   Gguf   License:other   Q2   Q4 0   Quantized   Qwen2   Region:us

Rank the Qwen1.5 110B Chat GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
Qwen1.5 110B Chat GGUF (second-state/Qwen1.5-110B-Chat-GGUF)

Best Alternatives to Qwen1.5 110B Chat GGUF

Best Alternatives
HF Rank
Context/RAM
Downloads
Likes
...n1.5 110B Chat 3.25bpw H6 EXL232K / 47.7 GB11
...n1.5 110B Chat 3.35bpw H6 EXL232K / 49 GB11
Qwen1.5 110B Chat 4bit32K / 62.2 GB194
Qwen1.5 110B Chat 8bit32K / 179.8 GB101
Qwen1.5 110B 4bit8K / 62.2 GB71
Qwen1.5 110B Chat GPTQ Int432K / 34 GB42
Qwen1.5 110B Chat AWQ32K / 61.1 GB80
Qwen1.5 110B Chat GPTQ Int432K / 61.7 GB3600612
Qwen1.5 110B Chat AWQ32K / 61.7 GB764695
Qwen1.5 110B Chat GPTQ Int832K / 115.9 GB130

Qwen1.5 110B Chat GGUF Parameters and Internals

LLM NameQwen1.5 110B Chat GGUF
RepositoryOpen on ๐Ÿค— 
Model NameQwen1.5-110B-Chat
Model CreatorQwen
Base Model(s)  Qwen1.5 110B Chat   Qwen/Qwen1.5-110B-Chat
Model Size110b
Required VRAM3.1 GB
Updated2024-05-20
Maintainersecond-state
Model Typeqwen2
Model Files  41.2 GB   32.2 GB: 1-of-2   21.5 GB: 2-of-2   32.1 GB: 1-of-2   30.8 GB: 2-of-2   32.0 GB: 1-of-3   32.1 GB: 2-of-3   3.1 GB: 3-of-3   32.1 GB: 1-of-2   31.3 GB: 2-of-2   32.0 GB: 1-of-3   32.1 GB: 2-of-3   12.5 GB: 3-of-3   32.1 GB: 1-of-3   32.0 GB: 2-of-3   14.8 GB: 3-of-3   31.9 GB: 1-of-3   32.0 GB: 2-of-3   27.3 GB: 3-of-3   32.1 GB: 1-of-4   31.9 GB: 2-of-4   32.2 GB: 3-of-4   22.0 GB: 4-of-4
Supported Languagesen
GGUF QuantizationYes
Quantization Typegguf|q2|q4_0|q4_k|q5_0|q5_k|q8_0
Model ArchitectureQwen2ForCausalLM
Licenseother
Context Length32768
Model Max Length32768
Transformers Version4.37.2
Vocabulary Size152064
Initializer Range0.02
Torch Data Typebfloat16

What open-source LLMs or SLMs are you in search of? 34817 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024042801