941 Downloads Updated 1 month ago
Updated 1 month ago
1 month ago
f8edd79ae2b9 · 5.9GB ·
Uploaded all but the Q8_0 as Unsloth DynamicQuant2 (DQ2) versions. These are Liquid AI’s on-device-first models using a hybrid Liquid architecture (gated convolutions + GQA) that’s 2x faster on CPU than standard transformers.
Temperature: A range of 0.4–0.6 works well for their instruction-following and reasoning tasks.
Liquid Foundation Models 2 (LFM2) — a family of fast, efficient models optimized for edge devices (phones, laptops, vehicles). The hybrid Liquid architecture provides superior speed and memory efficiency over traditional transformers.
The key difference between these models is their architecture and capacity:
Dense Models:
lfm2:350m: Entry-level dense model for ultra-low resource environments.
lfm2:700m: Compact dense model, outperforms Gemma 3 1B IT.
lfm2:1.2b: Dense model that competes with Qwen3-1.7B (47% larger).
lfm2:2.6b: Highest-quality dense model, outperforms 3B+ class models.
Mixture-of-Experts (On-device MoE):
Ideal for:
Fast on-device inference (CPU/NPU optimized)
Mobile AI applications and robotics
Low-latency chat and reasoning
English/Japanese-focused tasks with strong multilingual support