Phi 3 Mini 4K Instruct by microsoft

 ยป  All LLMs  ยป  microsoft  ยป  Phi 3 Mini 4K Instruct   URL Share it on

  Autotrain compatible   Code   Conversational   Custom code   En   Endpoints compatible   Fr   Instruct   Phi3   Region:us   Safetensors   Sharded   Tensorflow

Phi 3 Mini 4K Instruct Benchmarks

Phi 3 Mini 4K Instruct Parameters and Internals

Model Type 
text generation, language model
Use Cases 
Areas:
Research, Commercial applications
Applications:
General purpose AI systems, Computationally constrained environments
Primary Use Cases:
Memory/computational constraint scenarios, Latency bound applications, Tasks requiring mathematical and logical reasoning
Limitations:
Limited by language/data representation bias, Requires additional debiasing techniques for high-risk use cases
Considerations:
Evaluate performance and mitigate for safety and accuracy.
Additional Notes 
The model's performance improves when integrated with retrieval systems for external knowledge.
Supported Languages 
en (High proficiency), fr (Moderate proficiency)
Training Details 
Data Sources:
Publicly available documents, High-quality educational data, Newly created synthetic data, Chat format supervised data
Data Volume:
4.9 trillion tokens
Methodology:
Supervised fine-tuning and Direct Preference Optimization
Context Length:
4000
Training Time:
10 days
Hardware Used:
512 H100-80G GPUs
Model Architecture:
Dense decoder-only Transformer model
Safety Evaluation 
Methodologies:
Supervised fine-tuning, Direct Preference Optimization
Findings:
Strong reasoning capabilities, Improved instruction following
Risk Categories:
Misinformation, Bias
Ethical Considerations:
Use responsibly and ensure compliance with laws.
Responsible Ai Considerations 
Fairness:
Address bias through training data selection and filtering.
Transparency:
Encourage user feedback and continuous improvement.
Accountability:
Developers responsible for outputs and compliance.
Mitigation Strategies:
Use Retrieval Augmented Generation for grounding responses.
Input Output 
Input Format:
Chat format prompts
Accepted Modalities:
text
Output Format:
Text generation outputs
Performance Tips:
Utilize chat format for best results. Consider prompt engineering for improved performance.
Release Notes 
Version:
June 2024 Update
Notes:
Improved instruction following, structure output, and reasoning compared to the original release.
LLM NamePhi 3 Mini 4K Instruct
Repository ๐Ÿค—https://huggingface.co/microsoft/Phi-3-mini-4k-instruct 
Model Size3.8b
Required VRAM7.7 GB
Updated2024-11-21
Maintainermicrosoft
Model Typephi3
Instruction-BasedYes
Model Files  5.0 GB: 1-of-2   2.7 GB: 2-of-2
Supported Languagesen fr
Model ArchitecturePhi3ForCausalLM
Licensemit
Context Length4096
Model Max Length4096
Transformers Version4.40.2
Tokenizer ClassLlamaTokenizer
Padding Token<|endoftext|>
Vocabulary Size32064
Torch Data Typebfloat16
Phi 3 Mini 4K Instruct (microsoft/Phi-3-mini-4k-instruct)

Quantized Models of the Phi 3 Mini 4K Instruct

Model
Likes
Downloads
VRAM
Phi 3 Mini 4K Instruct GGUF35941 GB
...i 3 Mini 4K Instruct IMat GGUF13040 GB
...hi 3 Mini 4K Instruct Bnb 4bit3902 GB

Best Alternatives to Phi 3 Mini 4K Instruct

Best Alternatives
Context / RAM
Downloads
Likes
Phi 3.5 Mini Instruct128K / 7.7 GB651009640
Phi 3 Mini 128K Instruct128K / 7.7 GB7358311605
NuExtract 1.5128K / 7.7 GB107204113
NuExtract V1.5128K / 7.7 GB10851189
ECE EIFFEL 3Bv2128K / 7.7 GB50
Phi 3.5 Mini ITA128K / 7.7 GB788510
Flow Judge V0.1128K / 7.7 GB37945
Borea Phi 3.5 Mini Instruct Jp128K / 7.7 GB5299
Phi 3.5 Mini TitanFusion 0.1128K / 7.7 GB290
Borea Phi 3.5 Mini Instruct Jp128K / 7.7 GB2025

Rank the Phi 3 Mini 4K Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 38149 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241110