103 7 months ago

Using 4096 tokens for flash attention context window to work as intended. Trying a new template and system prompt to see how it reacts.

tools
ollama run mikepfunk28/deepseek8b_qwen3

Applications

Claude Code
Claude Code ollama launch claude --model mikepfunk28/deepseek8b_qwen3
Codex
Codex ollama launch codex --model mikepfunk28/deepseek8b_qwen3
OpenCode
OpenCode ollama launch opencode --model mikepfunk28/deepseek8b_qwen3
OpenClaw
OpenClaw ollama launch openclaw --model mikepfunk28/deepseek8b_qwen3

Models

View all →

Readme

No readme