Llava Phi3 by shtapm

 ยป  All LLMs  ยป  shtapm  ยป  Llava Phi3   URL Share it on

  Autotrain compatible   Conversational   Custom code   Endpoints compatible   Lora   Mipha phi   Region:us
Model Card on HF ๐Ÿค—: https://huggingface.co/shtapm/llava_phi3 
Llava Phi3 (shtapm/llava_phi3)

Llava Phi3 Parameters and Internals

LLM NameLlava Phi3
Repository ๐Ÿค—https://huggingface.co/shtapm/llava_phi3 
Required VRAM0.8 GB
Updated2024-07-04
Maintainershtapm
Model Files  0.8 GB   17.0 GB
Model ArchitectureAutoModelForCausalLM
Model Max Length2048
Is Biasednone
Tokenizer ClassLlamaTokenizer
Padding Token<unk>
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesmodel.layers.6.self_attn.qkv_proj|model.layers.16.mlp.down_proj|model.layers.22.self_attn.qkv_proj|model.layers.1.self_attn.qkv_proj|model.layers.9.self_attn.o_proj|model.layers.18.mlp.gate_up_proj|model.layers.26.self_attn.qkv_proj|model.layers.23.mlp.down_proj|model.layers.5.self_attn.o_proj|model.layers.6.mlp.down_proj|model.layers.2.mlp.gate_up_proj|model.layers.11.mlp.down_proj|model.layers.2.self_attn.qkv_proj|model.layers.27.self_attn.qkv_proj|model.layers.30.self_attn.o_proj|model.layers.17.mlp.gate_up_proj|model.layers.13.mlp.gate_up_proj|model.layers.6.mlp.gate_up_proj|model.layers.20.mlp.gate_up_proj|model.layers.27.mlp.gate_up_proj|model.layers.5.mlp.down_proj|model.layers.22.mlp.gate_up_proj|model.layers.26.self_attn.o_proj|model.layers.31.self_attn.qkv_proj|model.layers.7.mlp.down_proj|model.layers.21.mlp.down_proj|model.layers.16.mlp.gate_up_proj|model.layers.3.mlp.gate_up_proj|model.layers.11.mlp.gate_up_proj|model.layers.15.self_attn.o_proj|model.layers.4.mlp.gate_up_proj|model.layers.26.mlp.down_proj|model.layers.24.mlp.down_proj|model.layers.26.mlp.gate_up_proj|model.layers.28.self_attn.o_proj|model.layers.5.mlp.gate_up_proj|model.layers.31.mlp.down_proj|model.layers.27.mlp.down_proj|model.layers.16.self_attn.o_proj|model.layers.9.mlp.gate_up_proj|model.layers.23.mlp.gate_up_proj|model.layers.7.mlp.gate_up_proj|model.layers.28.mlp.down_proj|model.layers.23.self_attn.o_proj|model.layers.12.mlp.gate_up_proj|model.layers.1.mlp.gate_up_proj|model.layers.3.mlp.down_proj|model.layers.2.mlp.down_proj|model.layers.29.mlp.gate_up_proj|model.layers.7.self_attn.o_proj|model.layers.8.self_attn.qkv_proj|model.layers.14.mlp.gate_up_proj|model.layers.20.self_attn.qkv_proj|model.layers.3.self_attn.o_proj|model.layers.8.mlp.down_proj|model.layers.12.mlp.down_proj|model.layers.15.mlp.gate_up_proj|model.layers.24.mlp.gate_up_proj|model.layers.29.mlp.down_proj|model.layers.4.mlp.down_proj|model.layers.16.self_attn.qkv_proj|model.layers.12.self_attn.o_proj|model.layers.10.mlp.gate_up_proj|model.layers.24.self_attn.o_proj|model.layers.13.self_attn.qkv_proj|model.layers.17.self_attn.o_proj|model.layers.11.self_attn.qkv_proj|model.layers.22.self_attn.o_proj|model.layers.29.self_attn.qkv_proj|model.layers.23.self_attn.qkv_proj|model.layers.25.self_attn.qkv_proj|model.layers.22.mlp.down_proj|model.layers.19.self_attn.qkv_proj|model.layers.17.mlp.down_proj|model.layers.18.mlp.down_proj|model.layers.19.self_attn.o_proj|model.layers.25.mlp.down_proj|model.layers.30.self_attn.qkv_proj|model.layers.14.self_attn.o_proj|model.layers.10.self_attn.o_proj|model.layers.11.self_attn.o_proj|model.layers.5.self_attn.qkv_proj|model.layers.28.self_attn.qkv_proj|model.layers.12.self_attn.qkv_proj|model.layers.0.mlp.gate_up_proj|model.layers.20.self_attn.o_proj|model.layers.30.mlp.gate_up_proj|model.layers.21.self_attn.o_proj|model.layers.14.self_attn.qkv_proj|model.layers.7.self_attn.qkv_proj|model.layers.31.mlp.gate_up_proj|model.layers.3.self_attn.qkv_proj|model.layers.15.self_attn.qkv_proj|model.layers.18.self_attn.qkv_proj|model.layers.21.self_attn.qkv_proj|model.layers.25.mlp.gate_up_proj|model.layers.10.self_attn.qkv_proj|model.layers.4.self_attn.o_proj|model.layers.10.mlp.down_proj|model.layers.1.mlp.down_proj|model.layers.0.self_attn.o_proj|model.layers.28.mlp.gate_up_proj|model.layers.8.mlp.gate_up_proj|model.layers.13.self_attn.o_proj|model.layers.1.self_attn.o_proj|model.layers.24.self_attn.qkv_proj|model.layers.6.self_attn.o_proj|model.layers.9.self_attn.qkv_proj|model.layers.25.self_attn.o_proj|model.layers.30.mlp.down_proj|model.layers.17.self_attn.qkv_proj|model.layers.20.mlp.down_proj|model.layers.8.self_attn.o_proj|model.layers.9.mlp.down_proj|model.layers.29.self_attn.o_proj|model.layers.2.self_attn.o_proj|model.layers.4.self_attn.qkv_proj|model.layers.15.mlp.down_proj|model.layers.21.mlp.gate_up_proj|model.layers.27.self_attn.o_proj|model.layers.0.mlp.down_proj|model.layers.19.mlp.down_proj|model.layers.31.self_attn.o_proj|model.layers.0.self_attn.qkv_proj|model.layers.14.mlp.down_proj|model.layers.18.self_attn.o_proj|model.layers.13.mlp.down_proj|model.layers.19.mlp.gate_up_proj
LoRA Alpha256
LoRA Dropout0.05
R Param128

Best Alternatives to Llava Phi3

Best Alternatives
Context / RAM
Downloads
Likes
Fine Tune Sentimental Llama0K / 0 GB920
VLM2Vec LoRA0K / 0 GB1337
QuietStar Project0K /  GB42
Finetuned Llava Lora0K / 0.1 GB50
Alphace Email0K / 0.1 GB70
Qwen7B Haiguitang0K / 15.3 GB50
Accel0K / 0 GB120
Modelv30K / 13.5 GB50
Chinese Poetry Generation0K / 0 GB80
Partis Goodone0K / 16.1 GB31
Note: green Score (e.g. "73.2") means that the model is better than shtapm/llava_phi3.

Rank the Llava Phi3 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227