Mpt 30B Chat Q8 by Abzu

 ยป  All LLMs  ยป  Abzu  ยป  Mpt 30B Chat Q8   URL Share it on

  Arxiv:2010.04245   Arxiv:2108.12409   Arxiv:2205.14135   8-bit   Autotrain compatible   Codegen   Composer   Custom code   Instruct   Llm-foundry   Mosaicml   Mpt   Q8   Quantized   Region:us   Safetensors   Sharded   Tensorflow
Model Card on HF ๐Ÿค—: https://huggingface.co/Abzu/mpt-30b-chat-q8 

Mpt 30B Chat Q8 Benchmarks

Mpt 30B Chat Q8 (Abzu/mpt-30b-chat-q8)

Mpt 30B Chat Q8 Parameters and Internals

Model Type 
text generation, chatbot
Additional Notes 
The model was trained with MPT-30B tokenizer, based on EleutherAI/gpt-neox-20b tokenizer. The model employed techniques like FlashAttention and ALiBi for hardware efficiency.
Training Details 
Data Sources:
camel-ai/code, ehartford/wizard_vicuna_70k_unfiltered, anon8231489123/ShareGPT_Vicuna_unfiltered, teknium1/GPTeacher/roleplay-instruct-v2-final, teknium1/GPTeacher/codegen-isntruct, timdettmers/openassistant-guanaco, camel-ai/math, project-baize/baize-chatbot/medical_chat_data, project-baize/baize-chatbot/quora_chat_data, project-baize/baize-chatbot/stackoverflow_chat_data, camel-ai/biology, camel-ai/chemistry, camel-ai/ai_society, jondurbin/airoboros-gpt4-1.2, LongConversations, camel-ai/physics
Data Volume:
various public datasets
Methodology:
finetuning
Context Length:
8192
Training Time:
7.6 hours
Hardware Used:
64 H100 GPUs
Model Architecture:
modified decoder-only transformer
Input Output 
Input Format:
sequence of tokens (text)
Accepted Modalities:
text
Output Format:
text sequence (text generation)
Performance Tips:
Use 'trust_remote_code=True' for model architecture adaptations
LLM NameMpt 30B Chat Q8
Repository ๐Ÿค—https://huggingface.co/Abzu/mpt-30b-chat-q8 
Base Model(s)  Mpt 30B Chat   mosaicml/mpt-30b-chat
Model Size30b
Required VRAM30.4 GB
Updated2024-12-22
MaintainerAbzu
Model Typempt
Instruction-BasedYes
Model Files  10.0 GB: 1-of-4   9.9 GB: 2-of-4   9.9 GB: 3-of-4   0.6 GB: 4-of-4
Quantization Typeq8
Generates CodeYes
Model ArchitectureMPTForCausalLM
Licensecc-by-nc-sa-4.0
Model Max Length8192
Transformers Version4.30.2
Tokenizer ClassGPTNeoXTokenizer
Vocabulary Size50432
Torch Data Typebfloat16

Best Alternatives to Mpt 30B Chat Q8

Best Alternatives
Context / RAM
Downloads
Likes
Ct2fast Mpt 30B Chat0K / 30 GB102

Rank the Mpt 30B Chat Q8 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217