Beyonder 4x7B V2 by mlabonne

 ยป  All LLMs  ยป  mlabonne  ยป  Beyonder 4x7B V2   URL Share it on

  Autotrain compatible Beowolx/codeninja-1.0-openchat...   Conversational   Endpoints compatible Maywell/pivot-0.1-starling-lm-...   Merge   Mergekit   Mistral   Mixtral   Model-index   Moe   Openchat/openchat-3.5-1210   Region:us   Safetensors   Sharded   Tensorflow   Wizardlm/wizardmath-7b-v1.1

Beyonder 4x7B V2 Benchmarks

Beyonder 4x7B V2 (mlabonne/Beyonder-4x7B-v2)

Beyonder 4x7B V2 Parameters and Internals

Model Type 
text-generation
Additional Notes 
A Mixture of Experts model targeting different aspects such as chat, code, storytelling, and mathematics. Quantized models in GGUF, AWQ, GPTQ, EXL2 formats available.
Training Details 
Methodology:
Mixture of Experts created using mergekit (mixtral branch)
Context Length:
8000
LLM NameBeyonder 4x7B V2
Repository ๐Ÿค—https://huggingface.co/mlabonne/Beyonder-4x7B-v2 
Model Size24.2b
Required VRAM48.3 GB
Updated2025-02-22
Maintainermlabonne
Model Typemixtral
Model Files  9.9 GB: 1-of-5   10.0 GB: 2-of-5   10.0 GB: 3-of-5   10.0 GB: 4-of-5   8.4 GB: 5-of-5
Model ArchitectureMixtralForCausalLM
Licenseother
Context Length8192
Model Max Length8192
Transformers Version4.37.1
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32000
Torch Data Typebfloat16

Quantized Models of the Beyonder 4x7B V2

Model
Likes
Downloads
VRAM
Beyonder 4x7B V2 GGUF382738 GB
Beyonder 4x7B V2 GPTQ64712 GB
Beyonder 4x7B V2 AWQ31413 GB

Best Alternatives to Beyonder 4x7B V2

Best Alternatives
Context / RAM
Downloads
Likes
Dzakwan MoE 4x7b Beta32K / 48.4 GB38440
Beyonder 4x7B V332K / 48.3 GB394158
Calme 4x7B MoE V0.232K / 48.3 GB56362
Proto Athena 4x7B32K / 48.4 GB150
Proto Athena V0.2 4x7B32K / 48.4 GB80
Mera Mix 4x7B32K / 48.3 GB352518
Calme 4x7B MoE V0.132K / 48.3 GB39512
CognitiveFusion2 4x7B BF1632K / 48.3 GB36993
MixtureofMerges MoE 4x7b V532K / 48.3 GB19741
MixtureofMerges MoE 4x7b V432K / 48.3 GB19914
Note: green Score (e.g. "73.2") means that the model is better than mlabonne/Beyonder-4x7B-v2.

Rank the Beyonder 4x7B V2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43508 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227