3 yesterday

Kexity AI's first generation of flagship TLMs for efficient local inference.

tools thinking
ollama run KexityAI/kex

Applications

Claude Code
Claude Code ollama launch claude --model KexityAI/kex
OpenClaw
OpenClaw ollama launch openclaw --model KexityAI/kex
Hermes Agent
Hermes Agent ollama launch hermes --model KexityAI/kex
Codex
Codex ollama launch codex --model KexityAI/kex
OpenCode
OpenCode ollama launch opencode --model KexityAI/kex

Models

View all →

1 model

kex:latest

397MB · 40K context window · Text · yesterday

Readme

kexity.png

NOTE: Kex has been succeeded by Kex 1.5. We suggest using that instead.

Kex is Kexity AI’s first generation of flagship TLMs for efficient local inference. Kex supports tool calling, thinking, and features token-efficient thinking and reasoning for compute-constrained environments.

Use Case

This model is for customers with extremely constrained compute or low-latency applications. Kex punches above it’s weight in agentic use-cases, and is useful for tasks such as the following:

  • Agents running on edge/IoT devices with less than 512 MB of RAM
  • Low-latency chatbots and agents for environments where speed matters