318 Downloads Updated 11 months ago
ollama run sam860/phi4-mini:3.8b-Q4_K_S
Updated 11 months ago
11 months ago
fd12131f971f · 2.3GB ·
This model has been quantized from the fp16 model using ollama. I’ve provided it in Q4_K_S (default) and Q5_K_M. These quants were not uploaded by the ollama team in the main model tags, but I find them useful, so I wanted people to have easy access. I’ve also set these two parameters as default:
Temperature: 0.3 - this was on a whim
Context Length: 512 tokens - I’m in VRAM poverty and these phi models are only getting less and less efficient
Obviously you can change these settings easily by using a modelfile, or just setting the parameters manually in your UI of choice.
Phi-4-mini-instruct is a lightweight, open-source language model with 3.8 billion parameters, developed by Microsoft. It’s designed for broad commercial and research use, particularly in memory/compute-constrained environments and latency-bound scenarios, with a focus on strong reasoning capabilities (especially in math and logic).
This release incorporates user feedback from the Phi-3 series and introduces a new architecture for efficiency, a larger vocabulary for multilingual support, and improved post-training techniques for instruction following and function calling.