Phi 2 by microsoft

 ยป  All LLMs  ยป  microsoft  ยป  Phi 2   URL Share it on

  Autotrain compatible   Code   En   Endpoints compatible   Phi   Region:us   Safetensors   Sharded   Tensorflow
Model Card on HF ๐Ÿค—: https://huggingface.co/microsoft/phi-2 

Phi 2 Benchmarks

Phi 2 (microsoft/phi-2)

Phi 2 Parameters and Internals

Model Type 
text generation, code generation
Use Cases 
Areas:
research
Applications:
QA, chat, code format
Primary Use Cases:
QA format, chat format, code format
Limitations:
Generate Inaccurate Code and Facts, Limited Scope for code, Unreliable Responses to Instruction, Language Limitations, Potential Societal Biases, Toxicity, Verbosity
Additional Notes 
Phi-2 is intended for QA, chat, and code purposes. Model-generated text/code should be treated as a starting point. Users should be cautious when employing these models in applications.
Supported Languages 
en (Standard English)
Training Details 
Data Sources:
Phi-1.5, NLP synthetic texts, filtered websites
Data Volume:
250B tokens
Methodology:
Transformer-based model with next-word prediction objective
Context Length:
2048
Training Time:
14 days
Hardware Used:
96xA100-80G GPUs
Model Architecture:
Transformer-based model with next-word prediction objective
Input Output 
Accepted Modalities:
text
Performance Tips:
Phi-2 has an attention overflow issue (with FP16). If encountering this issue, enable/disable autocast on the PhiAttention.forward() function.
LLM NamePhi 2
Repository ๐Ÿค—https://huggingface.co/microsoft/phi-2 
Model Size2.8b
Required VRAM5.6 GB
Updated2025-02-05
Maintainermicrosoft
Model Typephi
Model Files  5.0 GB: 1-of-2   0.6 GB: 2-of-2
Supported Languagesen
Model ArchitecturePhiForCausalLM
Licensemit
Context Length2048
Model Max Length2048
Transformers Version4.37.0
Tokenizer ClassCodeGenTokenizer
Vocabulary Size51200
Torch Data Typefloat16

Quantized Models of the Phi 2

Model
Likes
Downloads
VRAM
Phi 2 GGUF19522003091 GB
Phi 2 GPTQ2914931 GB
Phi 2 GGUF63111 GB
Damysus 2.7B Chat GGUF0211 GB
Phi2 Bunny2185 GB
Phi 2 GGUF2861 GB

Best Alternatives to Phi 2

Best Alternatives
Context / RAM
Downloads
Likes
MFANN3bv0.24128K / 11.1 GB50
MFANN3b128K / 11.1 GB1160
MFANN3bv1.3128K / 11.1 GB130
MFANN3bv1.1128K / 11.1 GB160
MFANN3bv0.23128K / 11.1 GB60
MFANN3b SFT128K / 5.6 GB1690
MFANN3b Rebase128K / 11.1 GB100
MFANN3bv1.2126K / 11.1 GB320
MFANN Phigments Slerp V232K / 5.6 GB1340
MFANN3bv0.2232K / 11.1 GB50

Rank the Phi 2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227