Calme 4x7B MoE V0.1 by MaziyarPanahi

 ยป  All LLMs  ยป  MaziyarPanahi  ยป  Calme 4x7B MoE V0.1   URL Share it on

  7b   Autotrain compatible   Calme   Conversational   Generated from trainer   Mistral   Mixtral   Moe   Region:us   Safetensors   Sharded   Tensorflow

Calme 4x7B MoE V0.1 Benchmarks

Calme 4x7B MoE V0.1 (MaziyarPanahi/Calme-4x7B-MoE-v0.1)

Calme 4x7B MoE V0.1 Parameters and Internals

Model Type 
Mixture of Experts (MoE), text generation
Use Cases 
Areas:
text generation
Applications:
educational tools, academic research, commonsense reasoning, natural language applications, trustworthy information retrieval, math educational software, context understanding
Additional Notes 
Quantized versions available for use on computers without accelerated hardware.
Supported Languages 
English (high proficiency), French (high proficiency), Ukrainian (high proficiency)
Training Details 
Methodology:
Mixture of Experts, fine-tuning of four individidual Calme-7B models
Model Architecture:
Integrates four Calme-7B models with two experts per token
Input Output 
Accepted Modalities:
text
Output Format:
text
LLM NameCalme 4x7B MoE V0.1
Repository ๐Ÿค—https://huggingface.co/MaziyarPanahi/Calme-4x7B-MoE-v0.1 
Model NameCalme-4x7B-MoE-v0.1
Model CreatorMaziyarPanahi
Model Size24.2b
Required VRAM48.3 GB
Updated2025-03-13
MaintainerMaziyarPanahi
Model Typemixtral
Model Files  9.9 GB: 1-of-5   9.9 GB: 2-of-5   10.0 GB: 3-of-5   10.0 GB: 4-of-5   8.5 GB: 5-of-5
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.37.2
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32000
Torch Data Typebfloat16

Quantized Models of the Calme 4x7B MoE V0.1

Model
Likes
Downloads
VRAM
Calme 4x7B MoE V0.1 GGUF02518 GB

Best Alternatives to Calme 4x7B MoE V0.1

Best Alternatives
Context / RAM
Downloads
Likes
Dzakwan MoE 4x7b Beta32K / 48.4 GB20220
Proto Athena 4x7B32K / 48.4 GB150
Proto Athena V0.2 4x7B32K / 48.4 GB120
Calme 4x7B MoE V0.232K / 48.3 GB36532
Beyonder 4x7B V332K / 48.3 GB208158
Mera Mix 4x7B32K / 48.3 GB177118
CognitiveFusion2 4x7B BF1632K / 48.3 GB17793
...e 4x7B MoE ECE PRYMMAL Martial32K / 48.6 GB301
MixtureofMerges MoE 4x7b V532K / 48.3 GB1411
LCARS AI 1x4 003 SuperAI32K / 48.5 GB692
Note: green Score (e.g. "73.2") means that the model is better than MaziyarPanahi/Calme-4x7B-MoE-v0.1.

Rank the Calme 4x7B MoE V0.1 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 44950 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227