Dolphin 2.9.2 Phi 3 Medium by cognitivecomputations

 ยป  All LLMs  ยป  cognitivecomputations  ยป  Dolphin 2.9.2 Phi 3 Medium   URL Share it on

  Autotrain compatible Base model:finetune:unsloth/ph... Base model:unsloth/phi-3-mini-...   Conversational Dataset:cognitivecomputations/... Dataset:cognitivecomputations/... Dataset:cognitivecomputations/... Dataset:cognitivecomputations/...   Dataset:internlm/agent-flan Dataset:m-a-p/codefeedback-fil... Dataset:microsoft/orca-math-wo...   Dataset:teknium/openhermes-2.5   En   Endpoints compatible   Instruct   Mistral   Pytorch   Region:us   Sharded

Dolphin 2.9.2 Phi 3 Medium Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Dolphin 2.9.2 Phi 3 Medium (cognitivecomputations/dolphin-2.9.2-Phi-3-Medium)

Dolphin 2.9.2 Phi 3 Medium Parameters and Internals

Model Type 
instruction, conversational, coding
Additional Notes 
The model has been filtered to remove alignment and bias.
Supported Languages 
en (English), proficiency (Unknown)
Training Details 
Data Sources:
cognitivecomputations/Dolphin-2.9.2, teknium/OpenHermes-2.5, m-a-p/CodeFeedback-Filtered-Instruction, cognitivecomputations/dolphin-coder, cognitivecomputations/samantha-data, microsoft/orca-math-word-problems-200k, internlm/Agent-FLAN, cognitivecomputations/SystemChat-2.0
Methodology:
qLoRA fine-tuning with 4k sequence length
Context Length:
4000
Training Time:
3.5 days
Hardware Used:
8xL40S node provided by Crusoe Cloud
Model Architecture:
Based on Phi-3-Medium
Input Output 
Input Format:
ChatML prompt template
Accepted Modalities:
text
Output Format:
text outputs
Performance Tips:
Implement your own alignment layer before exposing the model as a service.
LLM NameDolphin 2.9.2 Phi 3 Medium
Repository ๐Ÿค—https://huggingface.co/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium 
Base Model(s)  Phi 3 Mini 4K Instruct   unsloth/Phi-3-mini-4k-instruct
Required VRAM28 GB
Updated2025-02-05
Maintainercognitivecomputations
Model Typemistral
Instruction-BasedYes
Model Files  4.9 GB: 1-of-6   5.0 GB: 2-of-6   4.9 GB: 3-of-6   5.0 GB: 4-of-6   5.0 GB: 5-of-6   3.2 GB: 6-of-6
Supported Languagesen
Model ArchitectureMistralForCausalLM
Licensemit
Context Length4096
Model Max Length4096
Transformers Version4.40.2
Tokenizer ClassLlamaTokenizer
Padding Token<|placeholder6|>
Vocabulary Size32064
Torch Data Typebfloat16

Quantized Models of the Dolphin 2.9.2 Phi 3 Medium

Model
Likes
Downloads
VRAM
... Phi 3 Medium AWQ 4bit Smashed0787 GB
... Phi 3 Medium Bnb 4bit Smashed0858 GB
... Phi 3 Medium HQQ 1bit Smashed062 GB
... Phi 3 Medium HQQ 4bit Smashed057 GB
... Phi 3 Medium HQQ 2bit Smashed054 GB

Best Alternatives to Dolphin 2.9.2 Phi 3 Medium

Best Alternatives
Context / RAM
Downloads
Likes
Krutrim 2 Instruct1000K / 49.3 GB1511
Mistral Instruct SCQA32K / 5 GB50
Lora Phi3 Mini4K / 5 GB60
...eZephir Sft Instruct Ead 16bit32K / 14.4 GB560
... Instruct Unsloth Merged 16bit4K / 7.6 GB11
... Mini 4K Instruct Bnb 4bit Ita4K / 7.6 GB27550
Phi 3 Mini Hospital Topic 504K / 7.6 GB790
MainPHI34K / 7.6 GB770
Model4K / 7.6 GB780
Phired4K / 7.6 GB120
Note: green Score (e.g. "73.2") means that the model is better than cognitivecomputations/dolphin-2.9.2-Phi-3-Medium.

Rank the Dolphin 2.9.2 Phi 3 Medium Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227