305 Downloads Updated 2 weeks ago
ollama run goekdenizguelmez/Gabliterated-Qwen3:thinking-4b-q4_k_m
Updated 2 weeks ago
2 weeks ago
4b4bcee4eaa3 · 2.5GB ·
With this model series, I introduce the first Gabliteration, a novel neural weight modification technique that advances beyond traditional abliteration methods through adaptive multi-directional projections with regularized layer selection. My new Gabliteration technique addresses the fundamental limitation of existing abliteration methods that compromise model quality while attempting to modify specific behavioral patterns. To understand the methods used behind Gabliteration, I suggest you to read the paper.
#P: 4
UGI: 32.25
W/10: 9.5
Writing: 11.3
NatInt: 16.67
Political lean: -26.0%
The Galbliterated version the worlds first 4B model with a W/10 benchmark of 9.5, proving the effectiveness of Gabliteration.
This series includes models ranging from 0.6B to 32B parameters, demonstrating the scalability and effectiveness of the Gabliteration technique across different model sizes.
Building upon the foundational work of Arditi et al. (2024) on single-direction abliteration, Gabliteration extends to a comprehensive multi-directional framework with theoretical guarantees. My method employs singular value decomposition on difference matrices between harmful and harmless prompt representations to extract multiple refusal directions.
If you use these models, please cite the original research:
Gülmez, G. (2025). Gabliteration: Adaptive Multi-Directional Neural Weight Modification for Selective Behavioral Alteration in Large Language Models. https://arxiv.org/abs/2512.18901
This model has reduced safety filtering and may generate sensitive or controversial outputs. Use responsibly and at your own risk.