TinyEnsemble 3x1.1B TinyMoE by JoPmt

 ยป  All LLMs  ยป  JoPmt  ยป  TinyEnsemble 3x1.1B TinyMoE   URL Share it on

78health/tinyllama 1.1b-functi...   Autotrain compatible Base model:cognitivecomputatio... Base model:daertml/tinygauss-1... Base model:merge:cognitivecomp... Base model:merge:daertml/tinyg... Cognitivecomputations/tinydolp...   Daertml/tinygauss-1.1b   Endpoints compatible   Frankenmoe   Lazymergekit   Merge   Mergekit   Mixtral   Moe   Region:us   Safetensors   Sharded   Tensorflow

TinyEnsemble 3x1.1B TinyMoE Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
TinyEnsemble 3x1.1B TinyMoE (JoPmt/TinyEnsemble-3x1.1B-TinyMoE)

TinyEnsemble 3x1.1B TinyMoE Parameters and Internals

Model Type 
Mixture of Experts (MoE)
Additional Notes 
The model is built using LazyMergekit technology, combining three different models. It employs positive prompts for routing inputs to the appropriate expert model within the ensemble.
LLM NameTinyEnsemble 3x1.1B TinyMoE
Repository ๐Ÿค—https://huggingface.co/JoPmt/TinyEnsemble-3x1.1B-TinyMoE 
Base Model(s)  cognitivecomputations/TinyDolphin-2.8-1.1b   78health/TinyLlama_1.1B-function-calling   DaertML/TinyGauss-1.1B   cognitivecomputations/TinyDolphin-2.8-1.1b   78health/TinyLlama_1.1B-function-calling   DaertML/TinyGauss-1.1B
Model Size2.6b
Required VRAM5.3 GB
Updated2025-02-22
MaintainerJoPmt
Model Typemixtral
Model Files  1.0 GB: 1-of-6   1.0 GB: 2-of-6   1.0 GB: 3-of-6   1.0 GB: 4-of-6   1.0 GB: 5-of-6   0.3 GB: 6-of-6
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.41.2
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32002
Torch Data Typebfloat16

Best Alternatives to TinyEnsemble 3x1.1B TinyMoE

Best Alternatives
Context / RAM
Downloads
Likes
TinyExperts V0 4x1B32K / 5.2 GB12980
...Oasst2 ChatML Cluster 3 2 1 V14K / 5.2 GB803
...x MoE Samantha Chatml Therabot4K / 5.2 GB151
LiteLlamix 8x460M 1T1K / 5.2 GB943
Note: green Score (e.g. "73.2") means that the model is better than JoPmt/TinyEnsemble-3x1.1B-TinyMoE.

Rank the TinyEnsemble 3x1.1B TinyMoE Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227