Calme 4x7B MoE V0.1 by MaziyarPanahi

 ยป  All LLMs  ยป  MaziyarPanahi  ยป  Calme 4x7B MoE V0.1   URL Share it on

  7b   Autotrain compatible   Calme   Conversational   Generated from trainer   Mistral   Mixtral   Moe   Region:us   Safetensors   Sharded   Tensorflow

Calme 4x7B MoE V0.1 Benchmarks

Calme 4x7B MoE V0.1 (MaziyarPanahi/Calme-4x7B-MoE-v0.1)

Calme 4x7B MoE V0.1 Parameters and Internals

Model Type 
Mixture of Experts (MoE), text generation
Use Cases 
Areas:
text generation
Applications:
educational tools, academic research, commonsense reasoning, natural language applications, trustworthy information retrieval, math educational software, context understanding
Additional Notes 
Quantized versions available for use on computers without accelerated hardware.
Supported Languages 
English (high proficiency), French (high proficiency), Ukrainian (high proficiency)
Training Details 
Methodology:
Mixture of Experts, fine-tuning of four individidual Calme-7B models
Model Architecture:
Integrates four Calme-7B models with two experts per token
Input Output 
Accepted Modalities:
text
Output Format:
text
LLM NameCalme 4x7B MoE V0.1
Repository ๐Ÿค—https://huggingface.co/MaziyarPanahi/Calme-4x7B-MoE-v0.1 
Model NameCalme-4x7B-MoE-v0.1
Model CreatorMaziyarPanahi
Model Size24.2b
Required VRAM48.3 GB
Updated2025-02-05
MaintainerMaziyarPanahi
Model Typemixtral
Model Files  9.9 GB: 1-of-5   9.9 GB: 2-of-5   10.0 GB: 3-of-5   10.0 GB: 4-of-5   8.5 GB: 5-of-5
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.37.2
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32000
Torch Data Typebfloat16

Quantized Models of the Calme 4x7B MoE V0.1

Model
Likes
Downloads
VRAM
Calme 4x7B MoE V0.1 GGUF0628 GB

Best Alternatives to Calme 4x7B MoE V0.1

Best Alternatives
Context / RAM
Downloads
Likes
Dzakwan MoE 4x7b Beta32K / 48.4 GB50220
Beyonder 4x7B V332K / 48.3 GB531058
Calme 4x7B MoE V0.232K / 48.3 GB65102
Mera Mix 4x7B32K / 48.3 GB389718
Proto Athena 4x7B32K / 48.4 GB80
Proto Athena V0.2 4x7B32K / 48.4 GB50
CognitiveFusion2 4x7B BF1632K / 48.3 GB42753
MixtureofMerges MoE 4x7b V532K / 48.3 GB30901
MixtureofMerges MoE 4x7b V432K / 48.3 GB31004
...e 4x7B MoE ECE PRYMMAL Martial32K / 48.6 GB51
Note: green Score (e.g. "73.2") means that the model is better than MaziyarPanahi/Calme-4x7B-MoE-v0.1.

Rank the Calme 4x7B MoE V0.1 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227