353 2 months ago

A model based on the GLM-4.6v-flash:9b q5_k_m, and uncensored. For local use I recommend editing the model context in modelfile as it is set to 128k. #EDIT: New local optimised model same with context 4096 https://ollama.com/ShreyanGondaliya/s5-reduced

ollama run ShreyanGondaliya/s5

Details

2 months ago

8309b927011d · 7.1GB ·

glm4
·
9.4B
·
Q5_K_M
[gMASK]<sop>{{ if .System }}<|system|> {{ .System }}{{ end }}{{ if .Prompt }}<|user|> {{ .Prompt }}{
You are an intelligent, advanced AI assistant with enhanced reasoning capabilities. You provide accu
{ "num_ctx": 128000, "num_gpu": 1, "num_thread": 8, "repeat_penalty": 1.2, "stop

Readme

A model based on the GLM-4.6v-flash:9b q5_k_m, and uncensored. For local use I recommend editing the model context in modelfile as it is set to 128k. Please note that a new model I have made with same weights but lowered context for same reasoning but local usage - https://ollama.com/ShreyanGondaliya/s5-reduced