197 Downloads Updated 7 months ago
DeepScaleR-1.5B-Preview is a fine-tuned language model based on DeepSeek-R1-Distilled-Qwen-1.5B, optimized for mathematical reasoning and long-context understanding. This guide provides instructions to run the model locally using Ollama.
If you haven’t installed Ollama, follow these steps:
Run the following command in your terminal:
curl -fsSL https://ollama.com/install.sh | sh
wsl --install
curl -fsSL https://ollama.com/install.sh | sh
For more details, check the official Ollama installation guide:
https://ollama.com/download
Once Ollama is installed, download DeepScaleR-1.5B-Preview:
ollama pull yasserrmd/DeepScaleR-1.5B-Preview
To start generating responses:
ollama run yasserrmd/DeepScaleR-1.5B-Preview
For interactive chat mode:
ollama chat yasserrmd/DeepScaleR-1.5B-Preview
You can modify the model behavior using a Modelfile.
Modelfile
and add the following content:FROM yasserrmd/DeepScaleR-1.5B-Preview
PARAMETER temperature 0.7
PARAMETER top_p 0.9
SYSTEM "You are an expert math problem solver. Provide detailed and accurate solutions."
ollama create my-deepscaler -f Modelfile
ollama run my-deepscaler
For improved inference speed:
- Run Ollama on a GPU-enabled system
- Use quantized versions (e.g., fp16
, int4
)
- Deploy on high-memory cloud instances (24GB RAM or more)
This model is released under the MIT License, promoting open-source AI development.
This model is built on DeepSeek-R1-Distilled-Qwen-1.5B, trained using GRPO at the Berkeley Sky Computing Lab and Berkeley AI Research.