Uni MoE V2 E2 by Uni-MoE

 ยป  All LLMs  ยป  Uni-MoE  ยป  Uni MoE V2 E2   URL Share it on

  Autotrain compatible   Endpoints compatible   Llava llama   Lora   Moe   Region:us
Model Card on HF ๐Ÿค—: https://huggingface.co/Uni-MoE/Uni-MoE-v2-e2 

Uni MoE V2 E2 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Uni MoE V2 E2 (Uni-MoE/Uni-MoE-v2-e2)

Uni MoE V2 E2 Parameters and Internals

LLM NameUni MoE V2 E2
Repository ๐Ÿค—https://huggingface.co/Uni-MoE/Uni-MoE-v2-e2 
Required VRAM0 GB
Updated2025-02-22
MaintainerUni-MoE
Model Files  0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.0 GB   0.3 GB   0.0 GB   0.0 GB
Model ArchitectureAutoModelForCausalLM
Licenseapache-2.0
Model Max Length2048
Is Biasednone
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesmodel.layers.17.self_attn.v_proj|model.layers.7.self_attn.o_proj|model.layers.26.self_attn.k_proj|model.layers.23.self_attn.o_proj|model.layers.28.self_attn.o_proj|model.layers.6.self_attn.q_proj|model.layers.22.self_attn.o_proj|model.layers.15.self_attn.o_proj|model.layers.29.self_attn.q_proj|model.layers.5.self_attn.v_proj|model.layers.16.self_attn.k_proj|model.layers.18.self_attn.k_proj|model.layers.3.self_attn.v_proj|model.layers.19.self_attn.v_proj|model.layers.7.self_attn.v_proj|model.layers.10.self_attn.v_proj|model.layers.3.self_attn.q_proj|model.layers.21.self_attn.q_proj|model.layers.31.self_attn.q_proj|model.layers.1.self_attn.v_proj|model.layers.13.self_attn.o_proj|model.layers.2.self_attn.k_proj|model.layers.12.self_attn.q_proj|model.layers.18.self_attn.o_proj|model.layers.24.self_attn.v_proj|model.layers.9.self_attn.o_proj|model.layers.11.self_attn.v_proj|model.layers.15.self_attn.q_proj|model.layers.8.self_attn.v_proj|model.layers.25.self_attn.o_proj|model.layers.24.self_attn.o_proj|model.layers.20.self_attn.v_proj|model.layers.27.self_attn.v_proj|model.layers.16.self_attn.v_proj|model.layers.10.self_attn.k_proj|model.layers.11.self_attn.q_proj|model.layers.17.self_attn.k_proj|model.layers.11.self_attn.o_proj|model.layers.4.self_attn.o_proj|model.layers.28.self_attn.k_proj|model.layers.23.self_attn.v_proj|model.layers.5.self_attn.o_proj|model.layers.8.self_attn.o_proj|model.layers.14.self_attn.v_proj|model.layers.14.self_attn.k_proj|model.layers.7.self_attn.q_proj|model.layers.1.self_attn.k_proj|model.layers.30.self_attn.v_proj|model.layers.22.self_attn.q_proj|model.layers.12.self_attn.k_proj|model.layers.25.self_attn.k_proj|model.layers.16.self_attn.q_proj|model.layers.25.self_attn.v_proj|model.layers.25.self_attn.q_proj|model.layers.19.self_attn.o_proj|model.layers.29.self_attn.k_proj|model.layers.3.self_attn.k_proj|model.layers.5.self_attn.k_proj|model.layers.6.self_attn.o_proj|model.layers.29.self_attn.o_proj|model.layers.9.self_attn.q_proj|model.layers.23.self_attn.k_proj|model.layers.4.self_attn.k_proj|model.layers.11.self_attn.k_proj|model.layers.17.self_attn.o_proj|model.layers.27.self_attn.k_proj|model.layers.1.self_attn.q_proj|model.layers.21.self_attn.v_proj|model.layers.12.self_attn.o_proj|model.layers.7.self_attn.k_proj|model.layers.6.self_attn.k_proj|model.layers.0.self_attn.q_proj|model.layers.0.self_attn.k_proj|model.layers.22.self_attn.k_proj|model.layers.15.self_attn.v_proj|model.layers.13.self_attn.k_proj|model.layers.3.self_attn.o_proj|model.layers.4.self_attn.q_proj|model.layers.5.self_attn.q_proj|model.layers.26.self_attn.q_proj|model.layers.16.self_attn.o_proj|model.layers.21.self_attn.o_proj|model.layers.13.self_attn.q_proj|model.layers.30.self_attn.k_proj|model.layers.31.self_attn.v_proj|model.layers.22.self_attn.v_proj|model.layers.31.self_attn.o_proj|model.layers.24.self_attn.q_proj|model.layers.28.self_attn.q_proj|model.layers.9.self_attn.v_proj|model.layers.1.self_attn.o_proj|model.layers.14.self_attn.q_proj|model.layers.29.self_attn.v_proj|model.layers.27.self_attn.q_proj|model.layers.13.self_attn.v_proj|model.layers.24.self_attn.k_proj|model.layers.2.self_attn.o_proj|model.layers.20.self_attn.o_proj|model.layers.27.self_attn.o_proj|model.layers.20.self_attn.k_proj|model.layers.2.self_attn.q_proj|model.layers.18.self_attn.q_proj|model.layers.6.self_attn.v_proj|model.layers.28.self_attn.v_proj|model.layers.17.self_attn.q_proj|model.layers.19.self_attn.q_proj|model.layers.0.self_attn.v_proj|model.layers.9.self_attn.k_proj|model.layers.8.self_attn.k_proj|model.layers.15.self_attn.k_proj|model.layers.26.self_attn.v_proj|model.layers.2.self_attn.v_proj|model.layers.21.self_attn.k_proj|model.layers.10.self_attn.q_proj|model.layers.30.self_attn.o_proj|model.layers.12.self_attn.v_proj|model.layers.14.self_attn.o_proj|model.layers.19.self_attn.k_proj|model.layers.10.self_attn.o_proj|model.layers.30.self_attn.q_proj|model.layers.31.self_attn.k_proj|model.layers.8.self_attn.q_proj|model.layers.18.self_attn.v_proj|model.layers.26.self_attn.o_proj|model.layers.20.self_attn.q_proj|model.layers.23.self_attn.q_proj|model.layers.4.self_attn.v_proj|model.layers.0.self_attn.o_proj
LoRA Alpha16
LoRA Dropout0.05
R Param8

Best Alternatives to Uni MoE V2 E2

Best Alternatives
Context / RAM
Downloads
Likes
Fine Tune Sentimental Llama0K / 0 GB920
VLM2Vec LoRA0K / 0 GB1337
QuietStar Project0K /  GB42
Finetuned Llava Lora0K / 0.1 GB50
Alphace Email0K / 0.1 GB70
Qwen7B Haiguitang0K / 15.3 GB50
Accel0K / 0 GB120
Modelv30K / 13.5 GB50
Chinese Poetry Generation0K / 0 GB80
Partis Goodone0K / 16.1 GB31
Note: green Score (e.g. "73.2") means that the model is better than Uni-MoE/Uni-MoE-v2-e2.

Rank the Uni MoE V2 E2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227