Yi 34B 200K RPMerge GPTQ by LoneStriker

 ยป  All LLMs  ยป  LoneStriker  ยป  Yi 34B 200K RPMerge GPTQ   URL Share it on

  Arxiv:2306.01708   Arxiv:2311.03099   4bit   Autotrain compatible   En   Endpoints compatible   Exl2   Exllama   Exllamav2   Gptq   Llama   Merge   Mergekit   Quantized   Region:us   Safetensors   Yi

Yi 34B 200K RPMerge GPTQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Yi 34B 200K RPMerge GPTQ (LoneStriker/Yi-34B-200K-RPMerge-GPTQ)

Yi 34B 200K RPMerge GPTQ Parameters and Internals

Model Type 
storytelling, instruction-following
Use Cases 
Areas:
storytelling, creative writing, role-playing
Applications:
novel writing, interactive fiction
Primary Use Cases:
>40K context, instruct-enhanced storytelling
Additional Notes 
Tested for novel-style continuation, assistant-type responses, and long context analysis without refusals. Specific to certain configurations for performance of storytelling in longer contexts.
Supported Languages 
en (proficient)
Training Details 
Data Sources:
DrNicefellow/ChatAllInOne-Yi-34B-200K-V1, migtissera/Tess-34B-v1.5b, cgato/Thespis-34b-v0.7, Doctor-Shotgun/limarpv3-yi-llama-34b-lora, adamo1139/yi-34b-200k-rawrr-dpo-2, migtissera/Tess-M-Creative-v1.0, NousResearch/Nous-Capybara-34B
Methodology:
Merge using DARE (Discrete Alignment-based Row Expansion) technique
Context Length:
40000
Input Output 
Input Format:
SYSTEM: {system_message} USER: {prompt} ASSISTANT:
Accepted Modalities:
text
Output Format:
text
Performance Tips:
Running Chinese models with large tokenizer vocabularies like Yi need careful parameter tuning due to large logit sampling tails.
LLM NameYi 34B 200K RPMerge GPTQ
Repository ๐Ÿค—https://huggingface.co/LoneStriker/Yi-34B-200K-RPMerge-GPTQ 
Model Size34b
Required VRAM21.2 GB
Updated2025-01-14
MaintainerLoneStriker
Model Typellama
Model Files  21.2 GB
Supported Languagesen
GPTQ QuantizationYes
Quantization Typeexl2|gptq|4bit
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length200000
Model Max Length200000
Transformers Version4.37.1
Tokenizer ClassLlamaTokenizer
Padding Token<unk>
Vocabulary Size64002
Torch Data Typefloat16

Best Alternatives to Yi 34B 200K RPMerge GPTQ

Best Alternatives
Context / RAM
Downloads
Likes
Smaug 34B V0.1 GPTQ195K / 21.2 GB111
Tess 34B V1.5B GPTQ195K / 18.6 GB317
...4B 200K DARE Megamerge V8 GPTQ195K / 18.6 GB233
Deepmoney 34B 200K Base GPTQ195K / 18.6 GB193
...y 34B 200K Chat Evaluator GPTQ195K / 18.6 GB133
...ous Capybara Limarpv3 34B GPTQ195K / 18.6 GB154
Bagel 34B V0.2 GPTQ195K / 18.6 GB322
Nontoxic Bagel 34B V0.2 GPTQ195K / 18.6 GB351
Bagel DPO 34B V0.2 GPTQ195K / 18.6 GB302
Yi 34B 200K AEZAKMI V2 GPTQ195K / 18.6 GB252
Note: green Score (e.g. "73.2") means that the model is better than LoneStriker/Yi-34B-200K-RPMerge-GPTQ.

Rank the Yi 34B 200K RPMerge GPTQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 41301 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227