Yi 34B 200K RPMerge AWQ by LoneStriker

 ยป  All LLMs  ยป  LoneStriker  ยป  Yi 34B 200K RPMerge AWQ   URL Share it on

  Arxiv:2306.01708   Arxiv:2311.03099   4-bit   Autotrain compatible   Awq   En   Endpoints compatible   Exl2   Exllama   Exllamav2   Llama   Merge   Mergekit   Quantized   Region:us   Safetensors   Sharded   Tensorflow   Yi

Yi 34B 200K RPMerge AWQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
๐ŸŒŸ Advertise your project ๐Ÿš€

Yi 34B 200K RPMerge AWQ Parameters and Internals

Model Type 
storytelling, instruction-following
Use Cases 
Areas:
storytelling, creative writing, role-playing
Applications:
novel writing, interactive fiction
Primary Use Cases:
>40K context, instruct-enhanced storytelling
Additional Notes 
Tested for novel-style continuation, assistant-type responses, and long context analysis without refusals. Specific to certain configurations for performance of storytelling in longer contexts.
Supported Languages 
en (proficient)
Training Details 
Data Sources:
DrNicefellow/ChatAllInOne-Yi-34B-200K-V1, migtissera/Tess-34B-v1.5b, cgato/Thespis-34b-v0.7, Doctor-Shotgun/limarpv3-yi-llama-34b-lora, adamo1139/yi-34b-200k-rawrr-dpo-2, migtissera/Tess-M-Creative-v1.0, NousResearch/Nous-Capybara-34B
Methodology:
Merge using DARE (Discrete Alignment-based Row Expansion) technique
Context Length:
40000
Input Output 
Input Format:
SYSTEM: {system_message} USER: {prompt} ASSISTANT:
Accepted Modalities:
text
Output Format:
text
Performance Tips:
Running Chinese models with large tokenizer vocabularies like Yi need careful parameter tuning due to large logit sampling tails.
LLM NameYi 34B 200K RPMerge AWQ
Repository ๐Ÿค—https://huggingface.co/LoneStriker/Yi-34B-200K-RPMerge-AWQ 
Model Size34b
Required VRAM19.2 GB
Updated2024-12-03
MaintainerLoneStriker
Model Typellama
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   4.3 GB: 4-of-4
Supported Languagesen
AWQ QuantizationYes
Quantization Typeexl2|awq
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length200000
Model Max Length200000
Transformers Version4.37.1
Tokenizer ClassLlamaTokenizer
Padding Token<unk>
Vocabulary Size64002
Torch Data Typefloat16
Yi 34B 200K RPMerge AWQ (LoneStriker/Yi-34B-200K-RPMerge-AWQ)

Best Alternatives to Yi 34B 200K RPMerge AWQ

Best Alternatives
Context / RAM
Downloads
Likes
Opus V1 34B AWQ195K / 19.2 GB91
Smaug 34B V0.1 AWQ195K / 19.2 GB42
Tess 34B V1.5B AWQ195K / 19.3 GB293
...34B 200K DARE Megamerge V8 AWQ195K / 19.3 GB252
...ey 34B 200K Chat Evaluator AWQ195K / 19.3 GB85
Deepmoney 34B 200K Base AWQ195K / 19.3 GB131
Nous Capybara Limarpv3 34B AWQ195K / 19.3 GB111
Bagel DPO 34B V0.2 AWQ195K / 19.3 GB77
Nontoxic Bagel 34B V0.2 AWQ195K / 19.3 GB122
Bagel 34B V0.2 AWQ195K / 19.3 GB112
Note: green Score (e.g. "73.2") means that the model is better than LoneStriker/Yi-34B-200K-RPMerge-AWQ.

Rank the Yi 34B 200K RPMerge AWQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 38770 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124