7 Downloads Updated 2 days ago
ollama run rockypod/svelte-coder
Updated 2 days ago
2 days ago
04fc5c27d630 · 9.0GB ·
A Svelte 5 / SvelteKit 2 specialist coding model in three sizes. Free to use under MIT. Built by rockypod on a homelab RTX 3090 Ti using continuous retrieval-augmented fine-tuning (RAFT).
# 14B — recommended default
ollama run rockypod/svelte-coder
# 8B — for hardware where 14B doesn't fit
ollama run rockypod/svelte-coder:v0.9.0-8b
# 4B — edge hardware
ollama run rockypod/svelte-coder:v0.9.0-4b
| Tag | Params | Disk | VRAM | When to pick |
|---|---|---|---|---|
:latest / :v0.9.0 |
14B | 8.4 GB | ~10 GB | Recommended. Best benchmark scores. |
:v0.9.0-8b |
8B | 5.0 GB | ~6 GB | Mid-tier GPUs, 16 GB VRAM laptops. |
:v0.9.0-4b |
4B | 2.5 GB | ~3 GB | Edge devices, entry-level GPUs. |
| Variant | 30Q spot | 204Q in-scope (rescored) |
|---|---|---|
| 14B (recommended) | 100% | 70.11% |
| 8B | 82.8% | 74.68% |
| 4B | 79.3% | 67.81% |
The 30Q spot exam is the cleaner instrument — pick by 30Q, not 204Q. The 204Q has known keyword-matching grader artifacts. The “8B beats 14B on 204Q” inversion is grader noise, not real capability — see the GitHub repo for the full transparency write-up.
$state, $derived, $effect, $props, $bindable+page.server.ts actions, load(), redirects, error
handling, route groups, hooksexport let, on:click,
<slot>) leak through more often on smaller variants. Use the 14B
for Svelte 4 conversion work.The Modelfile ships with these defaults — they’re load-bearing:
PARAMETER temperature 0.2
PARAMETER num_ctx 8192
PARAMETER num_predict 1500
PARAMETER repeat_penalty 1.5
Use a chat client that respects the Modelfile (Ollama CLI, Continue,
Zed, LM Studio with the included template). The OpenAI-compatible
/v1 endpoint silently drops num_ctx — use /api/chat if you need
to override context length over HTTP.