Phi-4 Mini 3.8B
by Microsoft · phi family
3.8B
parameters
text-generation code-generation reasoning math summarization
Phi-4 Mini 3.8B is Microsoft's compact reasoning model with 128K context, punching far above its weight on math and reasoning benchmarks. It continues the Phi tradition of proving that smaller, carefully trained models can compete with much larger ones. At Q4 it fits on any 8 GB GPU or Mac with minimal overhead — an excellent choice for users who want strong reasoning capabilities on resource-constrained hardware. Great for math, coding, and structured tasks.
Quick Start with Ollama
ollama run 3.8b-q4_K_M Quantization Options
| Format | File Size | VRAM Required | Quality | Ollama Tag |
|---|---|---|---|---|
| Q4_K_M rec | 2.5 GB | 4.5 GB | | 3.8b-q4_K_M |
| Q8_0 | 4.2 GB | 6.5 GB | | 3.8b-q8_0 |
| F16 | 7.8 GB | 10.5 GB | | 3.8b-fp16 |
Compatible Hardware
Q4_K_M requires 4.5 GB VRAM
Benchmark Scores
70.0
mmlu