85 Downloads Updated 7 months ago
DeepSeek-7B-1M is a hybrid model combining Qwen2.5-1.5B-1M and DeepSeek-R1-Distill-Qwen-1.5B, designed for enhanced reasoning, long-context understanding, and structured output generation. This model is optimized for mathematical problem-solving, code generation, and natural language understanding.
If Ollama is not installed, install it using:
curl -fsSL https://ollama.com/install.sh | sh
wsl --install
curl -fsSL https://ollama.com/install.sh | sh
For more details, check the official Ollama installation guide:
https://ollama.com/download
After installing Ollama, download DeepSeek-7B-1M:
ollama pull myrepo/deepseek-7b-1m
To start generating responses:
ollama run myrepo/deepseek-7b-1m
For interactive chat mode:
ollama chat myrepo/deepseek-7b-1m
Modelfile
You can adjust model behavior using a Modelfile.
Modelfile
and add the following:FROM myrepo/deepseek-7b-1m
PARAMETER temperature 0.7
PARAMETER top_p 0.9
SYSTEM "You are an AI expert trained for advanced reasoning, coding, and mathematical problem-solving. Provide detailed, structured, and optimized responses."
ollama create deepseek-7b-custom -f Modelfile
ollama run deepseek-7b-custom
For the best performance:
- Run Ollama on a GPU-enabled system
- Use quantized versions (e.g., fp16
, int4
) for efficient inference
- Deploy on high-memory cloud instances (32GB RAM or more)
This model is released under the MIT License, ensuring open and accessible AI development.
DeepSeek-7B-1M is built by merging Qwen2.5-1.5B-1M and DeepSeek-R1-Distill-Qwen-1.5B, leveraging their strengths in math, structured reasoning, and technical NLP.