97 Downloads Updated 2 months ago
A quantized version of Jan-Nano 4B, designed for deep research tasks, autonomous tool use, and integration with the Model Context Protocol (MCP). This model is optimized for use with Ollama and supports local agentic workflows.
ollama pull yasserrmd/jan-nano-4b
ollama run yasserrmd/jan-nano-4b
temperature
: 0.7
top_p
: 0.8
top_k
: 20
min_p
: 0.0
--hidethinking
to suppress self-verbalizationExample:
ollama run yasserrmd/jan-nano-4b \
--temperature 0.7 --top-p 0.8 --top-k 20 --hidethinking
โ MCP tool-calling (search, summarize, calculator, etc.) โ Optimized for 128K token context length โ Suitable for research, question answering, and autonomous agents โ Fast inference on consumer GPUs (8โ16GB VRAM)
Prompt:
User: /no_think
How has AI improved weather forecasting in the last five years?
Model Output: