KexityAI/ kex:latest

3 yesterday

Kexity AI's first generation of flagship TLMs for efficient local inference.

tools thinking
ollama run KexityAI/kex

Details

yesterday

c9901482cc8c · 397MB ·

qwen3
·
596M
·
Q4_K_M
{{- $lastUserIdx := -1 -}} {{- range $idx, $msg := .Messages -}} {{- if eq $msg.Role "user" }}{{ $la
{ "repeat_penalty": 1, "stop": [ "<|im_start|>", "<|im_end|>" ], "te

Readme

kexity.png

NOTE: Kex has been succeeded by Kex 1.5. We suggest using that instead.

Kex is Kexity AI’s first generation of flagship TLMs for efficient local inference. Kex supports tool calling, thinking, and features token-efficient thinking and reasoning for compute-constrained environments.

Use Case

This model is for customers with extremely constrained compute or low-latency applications. Kex punches above it’s weight in agentic use-cases, and is useful for tasks such as the following:

  • Agents running on edge/IoT devices with less than 512 MB of RAM
  • Low-latency chatbots and agents for environments where speed matters