MixtureofMerges MoE 4x7bRP V11 by jsfs11

 ยป  All LLMs  ยป  jsfs11  ยป  MixtureofMerges MoE 4x7bRP V11   URL Share it on

  Autotrain compatible Base model:biomistral/biomistr... Base model:chaoticneutrals/rp ... Base model:maziyarpanahi/mistr... Base model:merge:biomistral/bi... Base model:merge:chaoticneutra... Base model:merge:maziyarpanahi... Base model:merge:resplendentai... Base model:resplendentai/datur...   Biomistral/biomistral-dare-ns   Chaoticneutrals/rp vision 7b   Conversational   Endpoints compatible   Frankenmoe   Instruct   Lazymergekit Maziyarpanahi/mistral-7b-instr...   Merge   Mergekit   Mixtral   Moe   Region:us   Resplendentai/daturacookie 7b   Safetensors   Sharded   Tensorflow   Vision

MixtureofMerges MoE 4x7bRP V11 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
MixtureofMerges MoE 4x7bRP V11 (jsfs11/MixtureofMerges-MoE-4x7bRP-v11)

MixtureofMerges MoE 4x7bRP V11 Parameters and Internals

Model Type 
Mixture of Experts
Additional Notes 
The model combines multiple expert models to improve performance in text generation tasks and adapts outputs based on specific prompts.
Input Output 
Input Format:
Text
Accepted Modalities:
text
Output Format:
Text
Performance Tips:
Use appropriate prompts to tailor responses effectively.
LLM NameMixtureofMerges MoE 4x7bRP V11
Repository ๐Ÿค—https://huggingface.co/jsfs11/MixtureofMerges-MoE-4x7bRP-v11 
Base Model(s)  RP Vision 7B   DaturaCookie 7B   BioMistral DARE NS   MaziyarPanahi/Mistral-7B-Instruct-v0.3   ChaoticNeutrals/RP_Vision_7B   ResplendentAI/DaturaCookie_7B   BioMistral/BioMistral-DARE-NS   MaziyarPanahi/Mistral-7B-Instruct-v0.3
Model Size24.2b
Required VRAM48.3 GB
Updated2025-02-05
Maintainerjsfs11
Model Typemixtral
Instruction-BasedYes
Model Files  9.9 GB: 1-of-5   10.0 GB: 2-of-5   10.0 GB: 3-of-5   10.0 GB: 4-of-5   8.4 GB: 5-of-5
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.41.1
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32768
Torch Data Typebfloat16

Best Alternatives to MixtureofMerges MoE 4x7bRP V11

Best Alternatives
Context / RAM
Downloads
Likes
Proto Athena 4x7B32K / 48.4 GB80
Proto Athena V0.2 4x7B32K / 48.4 GB50
...erges MoE 4x7b V10 Mixtralv0.332K / 48.3 GB80
MoE Merging32K / 48.3 GB12140
Sixtyoneeighty 4x7B V132K / 48.3 GB520
...icon Mixtral87 Merged Torch21232K / 26.7 GB50
Eclipse Mistral 4x7b32K / 48.5 GB161
Kicon Mixtral87 Merged 4176632K / 26.7 GB50
Boundary Mistral 4x7b MoE32K / 48.7 GB1171
...x7b Quantized Portuguese Luana32K / 26.7 GB501
Note: green Score (e.g. "73.2") means that the model is better than jsfs11/MixtureofMerges-MoE-4x7bRP-v11.

Rank the MixtureofMerges MoE 4x7bRP V11 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227