148 2 months ago

Storytelling mixture of experts model for consumer GPUs. Made by DavidAU (Huggingface).

tools
ollama run LESSTHANSUPER/DARK_PLANET_REBEL_FURY-Llama3-25b:Q3_K_S

Applications

Claude Code
Claude Code ollama launch claude --model LESSTHANSUPER/DARK_PLANET_REBEL_FURY-Llama3-25b:Q3_K_S
Codex
Codex ollama launch codex --model LESSTHANSUPER/DARK_PLANET_REBEL_FURY-Llama3-25b:Q3_K_S
OpenCode
OpenCode ollama launch opencode --model LESSTHANSUPER/DARK_PLANET_REBEL_FURY-Llama3-25b:Q3_K_S
OpenClaw
OpenClaw ollama launch openclaw --model LESSTHANSUPER/DARK_PLANET_REBEL_FURY-Llama3-25b:Q3_K_S

Models

View all →

Readme

DARK PLANET REBEL FURY / I-MATRIX / 25B (4X8B) / I-QUANT

A more recent model from DavidAU’s “Dark Planet” line, one the creator favored themselves among their similar MoE (mixture of experts) models. This model was uploaded for the speed MoEs provide relative for their size, and as I have had good experience with the Dark Planet series. If a higher active parameter model is preferred, there is a 13-billion active parameter Mixtral model available. To stuff as many parameters in as little VRAM as possible, weighted K and I-quants will be listed.

Note that I-quants forfeit some token generation speed relative to K-quants in exchange for storage efficiency. Either of the 4-bit quantizations are recommended for 16GB GPUs. These models were taken from GGUF formats from Huggingface.

Original model (DavidAU):

GGUF weighted quantizations (mradermacher):

[No obligatory model picture. Ollama did not like it.]