TinyEnsemble 3x1.1B TinyMoE by JoPmt

 ยป  All LLMs  ยป  JoPmt  ยป  TinyEnsemble 3x1.1B TinyMoE   URL Share it on

78health/tinyllama 1.1b-functi...   Autotrain compatible Base model:78health/tinyllama ... Base model:cognitivecomputatio... Base model:daertml/tinygauss-1... Base model:merge:78health/tiny... Base model:merge:cognitivecomp... Base model:merge:daertml/tinyg... Cognitivecomputations/tinydolp...   Daertml/tinygauss-1.1b   Endpoints compatible   Frankenmoe   Lazymergekit   Merge   Mergekit   Mixtral   Moe   Region:us   Safetensors   Sharded   Tensorflow

TinyEnsemble 3x1.1B TinyMoE Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

TinyEnsemble 3x1.1B TinyMoE Parameters and Internals

LLM NameTinyEnsemble 3x1.1B TinyMoE
Repository ๐Ÿค—https://huggingface.co/JoPmt/TinyEnsemble-3x1.1B-TinyMoE 
Base Model(s)  cognitivecomputations/TinyDolphin-2.8-1.1b   78health/TinyLlama_1.1B-function-calling   DaertML/TinyGauss-1.1B   cognitivecomputations/TinyDolphin-2.8-1.1b   78health/TinyLlama_1.1B-function-calling   DaertML/TinyGauss-1.1B
Model Size2.6b
Required VRAM5.3 GB
Updated2024-09-07
MaintainerJoPmt
Model Typemixtral
Model Files  1.0 GB: 1-of-6   1.0 GB: 2-of-6   1.0 GB: 3-of-6   1.0 GB: 4-of-6   1.0 GB: 5-of-6   0.3 GB: 6-of-6
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.41.2
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32002
Torch Data Typebfloat16
TinyEnsemble 3x1.1B TinyMoE (JoPmt/TinyEnsemble-3x1.1B-TinyMoE)

Best Alternatives to TinyEnsemble 3x1.1B TinyMoE

Best Alternatives
Context / RAM
Downloads
Likes
TinyExperts V0 4x1B32K / 5.2 GB6780
...x MoE Samantha Chatml Therabot4K / 5.2 GB101
...Oasst2 ChatML Cluster 3 2 1 V14K / 5.2 GB113
LiteLlamix 8x460M 1T1K / 5.2 GB213
Note: green Score (e.g. "73.2") means that the model is better than JoPmt/TinyEnsemble-3x1.1B-TinyMoE.

Rank the TinyEnsemble 3x1.1B TinyMoE Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 35693 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024072803