TinyMistral 248Mx4 MoE Not Tuned Pls Help by gate369

 ยป  All LLMs  ยป  gate369  ยป  TinyMistral 248Mx4 MoE Not Tuned Pls Help   URL Share it on

  Autotrain compatible Base model:jtatman/tinymistral... Base model:locutusque/tinymist... Base model:locutusque/tinymist... Base model:m4-ai/tinymistral-2... Base model:merge:jtatman/tinym... Base model:merge:locutusque/ti... Base model:merge:locutusque/ti... Base model:merge:m4-ai/tinymis...   Endpoints compatible   Frankenmoe   Instruct Jtatman/tinymistral-v2-pycoder...   Lazymergekit Locutusque/tinymistral-248m-in... Locutusque/tinymistral-248m-v2... M4-ai/tinymistral-248m-v2-clea...   Merge   Mergekit   Mixtral   Model-index   Moe   Region:us   Safetensors   Sharded   Tensorflow

TinyMistral 248Mx4 MoE Not Tuned Pls Help Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

TinyMistral 248Mx4 MoE Not Tuned Pls Help Parameters and Internals

LLM NameTinyMistral 248Mx4 MoE Not Tuned Pls Help
Repository ๐Ÿค—https://huggingface.co/gate369/TinyMistral-248Mx4-MOE-not-tuned-pls-help 
Base Model(s)  M4-ai/TinyMistral-248M-v2-cleaner   TinyMistral 248M Instruct   jtatman/tinymistral-v2-pycoder-instuct-248m   TinyMistral 248M V2 Instruct   M4-ai/TinyMistral-248M-v2-cleaner   Locutusque/TinyMistral-248M-Instruct   jtatman/tinymistral-v2-pycoder-instuct-248m   Locutusque/TinyMistral-248M-v2-Instruct
Model Size701.1m
Required VRAM1.4 GB
Updated2024-09-07
Maintainergate369
Model Typemixtral
Instruction-BasedYes
Model Files  1.4 GB: 1-of-1
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length32768
Model Max Length32768
Transformers Version4.36.2
Tokenizer ClassLlamaTokenizer
Padding Token<|bos|>
Vocabulary Size32005
Torch Data Typefloat16
TinyMistral 248Mx4 MoE Not Tuned Pls Help (gate369/TinyMistral-248Mx4-MOE-not-tuned-pls-help)

Rank the TinyMistral 248Mx4 MoE Not Tuned Pls Help Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 35693 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024072803