19.4K 4 days ago

Specialized uncensored/abliterated quants for new OpenAI 20B MOE - Mixture of Experts Model at 80+ T/S (quantized Q5_1)

thinking 20b

4 days ago

dfee3e8688f3 · 16GB

gpt-oss
·
20.9B
·
Q5_1
<|start|>system<|message|>You are ChatGPT, a large language model trained by OpenAI. Knowledge cutof
{ "temperature": 1 }

Readme

Based on https://huggingface.co/DavidAU/OpenAi-GPT-oss-20b-abliterated-uncensored-NEO-Imatrix-gguf

Feature Value
vision false
thinking true (without switching off)
tools not working
Device Speed Version
RTX 3090 24gb ~159 token/s !! Q5_1
M1 Max 32gb ~58 token/s Q5_1
i7-12700H + RTX 3070ti Mobile 8gb (55%/45% CPU/GPU) ~17 token/s Q5_1