160 Downloads Updated 3 days ago
ollama run studiobrn/uncensoredmodAI
Updated 3 days ago
3 days ago
f12ec0a1f258 · 18GB ·
uncensoredmodAI is a direct-response Ollama model profile designed for coding, technical assistance, automation workflows, and practical local AI usage.
This model is optimized for users who want clear final answers, complete code output, and less unnecessary explanation.
--think=false for direct answersollama run studiobrn/uncensoredmodai
For direct final answers without visible thinking output:
ollama run studiobrn/uncensoredmodai --think=false
For smoother usage on lower-memory machines, especially Apple Silicon / 16GB RAM systems:
OLLAMA_CONTEXT_LENGTH=4096 \
OLLAMA_NUM_PARALLEL=1 \
OLLAMA_MAX_LOADED_MODELS=1 \
OLLAMA_FLASH_ATTENTION=1 \
OLLAMA_KV_CACHE_TYPE=q4_0 \
OLLAMA_KEEP_ALIVE=1m \
ollama serve
Then run the model in another terminal:
ollama run studiobrn/uncensoredmodai --think=false
Use direct prompts like:
Do not output thinking.
Do not output reasoning traces.
Only provide the final answer.
Write complete code.
Do not stop in the middle of a function, class, file, JSON, or command.
Example:
Write a complete Node.js Express API in one file. Only return the code.
This is a large 25.8B model. On 16GB Apple Silicon machines, it may use partial CPU/GPU offloading, so response speed can vary.
For best results:
--think=falsenum_predict when longer code output is neededcurl http://localhost:11434/api/chat -d '{
"model": "studiobrn/uncensoredmodai",
"messages": [
{
"role": "user",
"content": "Write a clean Python FastAPI example. Only return the code."
}
],
"think": false,
"stream": true,
"options": {
"num_ctx": 4096,
"num_predict": 2048,
"temperature": 0.1
}
}'
uncensoredmodAI is built for direct, practical, uncensored, and productive local AI workflows.
The main goal is simple:
Fully uncensored local AI. Less unnecessary thinking output. More complete final answers. Better assistance.