921 Downloads Updated 1 month ago
Name
17 models
LFM2:350m
379MB · 125K context window · Text · 1 month ago
LFM2:700m
792MB · 125K context window · Text · 1 month ago
LFM2:1.2b
843MB · 32K context window · Text · 1 month ago
LFM2:2.6b
1.8GB · 125K context window · Text · 1 month ago
LFM2:8b
5.9GB · 125K context window · Text · 1 month ago
Uploaded all but the Q8_0 as Unsloth DynamicQuant2 (DQ2) versions. These are Liquid AI’s on-device-first models using a hybrid Liquid architecture (gated convolutions + GQA) that’s 2x faster on CPU than standard transformers.
Temperature: A range of 0.4–0.6 works well for their instruction-following and reasoning tasks.
Liquid Foundation Models 2 (LFM2) — a family of fast, efficient models optimized for edge devices (phones, laptops, vehicles). The hybrid Liquid architecture provides superior speed and memory efficiency over traditional transformers.
The key difference between these models is their architecture and capacity:
Dense Models:
lfm2:350m: Entry-level dense model for ultra-low resource environments.
lfm2:700m: Compact dense model, outperforms Gemma 3 1B IT.
lfm2:1.2b: Dense model that competes with Qwen3-1.7B (47% larger).
lfm2:2.6b: Highest-quality dense model, outperforms 3B+ class models.
Mixture-of-Experts (On-device MoE):
Ideal for:
Fast on-device inference (CPU/NPU optimized)
Mobile AI applications and robotics
Low-latency chat and reasoning
English/Japanese-focused tasks with strong multilingual support