Models
Docs
Pricing
Sign in
Download
Models
Download
Docs
Pricing
Sign in
mikepfunk28
/
deepseek8b_qwen3
103
Downloads
Updated
7 months ago
Using 4096 tokens for flash attention context window to work as intended. Trying a new template and system prompt to see how it reacts.
Using 4096 tokens for flash attention context window to work as intended. Trying a new template and system prompt to see how it reacts.
Cancel
tools
Name
1 model
Size
Context
Input
deepseek8b_qwen3:latest
77299afe1734
• 5.2GB • 128K context window •
Text input • 7 months ago
Text input • 7 months ago
deepseek8b_qwen3:latest
5.2GB
128K
Text
77299afe1734
· 7 months ago