Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed by PrunaAI

 ยป  All LLMs  ยป  PrunaAI  ยป  Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed   URL Share it on

  4bit   Autotrain compatible Base model:microsoft/phi-3-min...   Conversational   Custom code   Endpoints compatible   Instruct   Phi3   Pruna-ai   Quantized   Region:us

Rank the Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed (PrunaAI/microsoft-Phi-3-mini-128k-instruct-HQQ-4bit-smashed)

Best Alternatives to Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed

Best Alternatives
HF Rank
Context/RAM
Downloads
Likes
...128K Instruct HQQ 2bit Smashed128K / 1.4 GB240
...28K Instruct Ov Fp16 Int4 Asym128K / 2.5 GB760
...m 128K Instruct 3.0bpw H6 EXL2128K / 5.6 GB70
...m 128K Instruct 5.0bpw H6 EXL2128K / 8.9 GB150
...m 128K Instruct 6.0bpw H6 EXL2128K / 10.7 GB43
...m 128K Instruct 8.0bpw H8 EXL2128K / 13.4 GB204
...dium 128K Instruct 8 0bpw EXL2128K / 13.4 GB61
Phi 3 Mini 4K Instruct Fp164K /  GB81
...i 4K Instruct HQQ 1bit Smashed4K / 0.9 GB650
...Instruct V0.3 HQQ 1bit Smashed4K / 0.9 GB80

Microsoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed Parameters and Internals

LLM NameMicrosoft Phi 3 Mini 128K Instruct HQQ 4bit Smashed
RepositoryOpen on ๐Ÿค— 
Base Model(s)  Phi 3 Mini 128K Instruct   microsoft/Phi-3-mini-128k-instruct
Required VRAM2.3 GB
Updated2024-07-07
MaintainerPrunaAI
Model Typephi3
Instruction-BasedYes
Model Files  2.3 GB
Quantization Type4bit
Model ArchitecturePhi3ForCausalLM
Context Length131072
Model Max Length131072
Transformers Version4.40.0
Tokenizer ClassLlamaTokenizer
Padding Token<|endoftext|>
Vocabulary Size32064
Initializer Range0.02
Torch Data Typebfloat16
Embedding Dropout0

What open-source LLMs or SLMs are you in search of? 34531 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024042801