LLM Explorer: A Curated Large Language Model Directory and Analytics  // 

Mpt 30B Peft Compatible by eluzhnica

What open-source LLMs or SLMs are you in search of? 18732 in total.

 ยป  All LLMs  ยป  eluzhnica  ยป  Mpt 30B Peft Compatible   URL Share it on

  Arxiv:1909.08053   Arxiv:2010.04245   Arxiv:2108.12409   Arxiv:2205.14135   Arxiv:2302.06675   Arxiv:2302.13971   Autotrain compatible   Composer   Custom code   Dataset:allenai/c4   Dataset:allenai/s2orc Dataset:bigcode/the-stack-dedu...   Dataset:mc4 Dataset:togethercomputer/redpa...   License:apache-2.0   Llm-foundry   Mosaicml   Mpt   Pytorch   Region:us   Sharded   Streamingdatasets

Rank the Mpt 30B Peft Compatible Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
Mpt 30B Peft Compatible (eluzhnica/mpt-30b-peft-compatible)

Best Alternatives to Mpt 30B Peft Compatible

Best Alternatives
HF Rank
Mpt 30B Chat55.380K / 60.1 GB12363192
Mpt 30B Instruct53.40K / 60.1 GB935396
Mpt 30B52.770K / 60.1 GB46681331
Mpt 30B Chat Q848.60K / 30.4 GB31
Mpt 30B Instruct Q846.80K / 30.4 GB35
Mpt 30B Qlora Multi GPU0K /  GB31
Ct2fast Mpt 30B Instruct0K / 30 GB24
Ct2fast Mpt 30B0K / 30 GB23
Ct2fast Mpt 30B Chat0K / 30 GB22
Mpt 30B Qlora Compatible0K / 60.1 GB111
Note: green Score (e.g. "73.2") means that the model is better than eluzhnica/mpt-30b-peft-compatible.

Mpt 30B Peft Compatible Parameters and Internals

LLM NameMpt 30B Peft Compatible
RepositoryOpen on ๐Ÿค— 
Model Size30b
Required VRAM60.1 GB
Model Typempt
Model Files  9.8 GB: 1-of-7   9.9 GB: 2-of-7   9.9 GB: 3-of-7   9.9 GB: 4-of-7   9.9 GB: 5-of-7   9.9 GB: 6-of-7   0.8 GB: 7-of-7
Model ArchitectureMPTForCausalLM
Model Max Length8192
Transformers Version4.28.1
Tokenizer ClassGPTNeoXTokenizer
Vocabulary Size50432
Torch Data Typebfloat16
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024022003