Updated 12 hours ago
ollama run aisingapore/Apertus-SEA-LION-v4-8B-IT
ollama launch claude --model aisingapore/Apertus-SEA-LION-v4-8B-IT
ollama launch codex --model aisingapore/Apertus-SEA-LION-v4-8B-IT
ollama launch opencode --model aisingapore/Apertus-SEA-LION-v4-8B-IT
ollama launch openclaw --model aisingapore/Apertus-SEA-LION-v4-8B-IT
Name
5 models
Apertus-SEA-LION-v4-8B-IT:latest
5.1GB · 64K context window · Text · 12 hours ago
Apertus-SEA-LION-v4-8B-IT:q4_k_m
5.1GB · 64K context window · Text · 12 hours ago
Apertus-SEA-LION-v4-8B-IT:q6_k
6.6GB · 64K context window · Text · 13 hours ago
Apertus-SEA-LION-v4-8B-IT:q8_0
8.6GB · 64K context window · Text · 14 hours ago
Apertus-SEA-LION-v4-8B-IT:f16
16GB · 64K context window · Text · 21 hours ago
[Last update: 2026-02-05]
SEA-LION is a collection of Large Language Models (LLMs) which have been pretrained and instruct-tuned for the Southeast Asia (SEA) region.
Apertus-SEA-LION-v4-8B-IT is a 8-billion parameter model built upon the Apertus-8B-Instruct architecture. To ensure domain adaptation for the region, the model underwent rigorous post-training on a curated dataset of approximately 6.4 million instruction-text pairs.
This extensive post-training instills multilingual and multicultural fluency, covering key SEA languages such as Indonesian, Vietnamese, Thai, Filipino, Tamil, Burmese, Malay. This curated dataset also includes a filtered open sourced set of tool-calling instruction-text pairs to impart these capabilities, in addition to linguistic fluency.
Apertus-SEA-LION-v4-8B-IT is designed as a fully open model to align with this core philosophy, we have released the datasets used for post-training, as well as the evaluation codes and datasets used to evaluate the model.
These resources can be accessed via the link below.
SEA-LION stands for Southeast Asian Languages In One Network.
We performed Post-Training in English and SEA languages on Apertus-8B-Instruct-2509, a decoder model using the Apertus architecture, to create Apertus-SEA-LION-v4-8B-IT.
For tokenization, the model employs the default tokenizer used in Apertus-8B-Instruct-2509.