Saiga Phi3 Medium Sft M1 D2 Kto M5 D7 by IlyaGusev

 ยป  All LLMs  ยป  IlyaGusev  ยป  Saiga Phi3 Medium Sft M1 D2 Kto M5 D7   URL Share it on

  Arxiv:1910.09700   Mistral   Peft   Region:us   Safetensors   Sharded   Tensorflow

Saiga Phi3 Medium Sft M1 D2 Kto M5 D7 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Saiga Phi3 Medium Sft M1 D2 Kto M5 D7 (IlyaGusev/saiga_phi3_medium_sft_m1_d2_kto_m5_d7)

Saiga Phi3 Medium Sft M1 D2 Kto M5 D7 Parameters and Internals

LLM NameSaiga Phi3 Medium Sft M1 D2 Kto M5 D7
Repository ๐Ÿค—https://huggingface.co/IlyaGusev/saiga_phi3_medium_sft_m1_d2_kto_m5_d7 
Base Model(s)  models/saiga_phi3_medium_sft_m1_d2   models/saiga_phi3_medium_sft_m1_d2
Model Size14b
Required VRAM28 GB
Updated2025-02-22
MaintainerIlyaGusev
Model Typemistral
Model Files  4.9 GB: 1-of-6   5.0 GB: 2-of-6   4.9 GB: 3-of-6   5.0 GB: 4-of-6   5.0 GB: 5-of-6   3.2 GB: 6-of-6
Model ArchitectureMistralForCausalLM
Context Length4096
Model Max Length4096
Transformers Version4.42.0.dev0
Tokenizer ClassLlamaTokenizer
Padding Token<s>
Vocabulary Size32064
Torch Data Typebfloat16

Best Alternatives to Saiga Phi3 Medium Sft M1 D2 Kto M5 D7

Best Alternatives
Context / RAM
Downloads
Likes
...ral Nemo Instruct 14B Merge V11000K / 24.6 GB190
K2S3 14B V0.232K / 28.7 GB280
Wendigo 14B Alpha432K / 28.4 GB12880
Qwen1.5 14B Chat Mistral32K / 28.6 GB202
Mistral 14B Merge Base32K / 28.4 GB20062
Synthetic Minstrel 14B32K / 27.6 GB223
Wandering Minstrel 14B32K / 27.6 GB113
Barcenas 14B Phi 3 Medium ORPO4K / 28 GB56435
SauerkrautLM Phi 3 Medium4K / 28 GB55519
...2.9.2 Phi 3 Medium Abliterated4K / 28 GB393517
Note: green Score (e.g. "73.2") means that the model is better than IlyaGusev/saiga_phi3_medium_sft_m1_d2_kto_m5_d7.

Rank the Saiga Phi3 Medium Sft M1 D2 Kto M5 D7 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227