4,544 Downloads Updated 1 month ago
ollama run sam860/LFM2:2.6b-exp-Q8_0
Updated 1 month ago
1 month ago
a1d8051fa86f · 2.7GB ·
Uploaded as Unsloth DynamicQuant2 (DQ2) versions. These are Liquid AI’s on-device-first models using a hybrid Liquid architecture (gated convolutions + GQA) that’s 2x faster on CPU than standard transformers.
All models here are Q4_0. The unsloth dq2 method may preserve most of the performance of the q8 model, but doesn’t guarantee it. Do your own testing to make sure it really works for your use case.
Temperature: A range of 0.4–0.6 works well for their instruction-following and reasoning tasks.
Liquid Foundation Models 2 (LFM2) — a family of fast, efficient models optimized for edge devices (phones, laptops, vehicles). The hybrid Liquid architecture provides superior speed and memory efficiency over traditional transformers.
The key difference between these models is their architecture and capacity:
Dense Models:
lfm2:350m: Entry-level dense model for ultra-low resource environments.
lfm2:700m: Compact dense model, outperforms Gemma 3 1B IT.
lfm2:1.2b: Dense model that competes with Qwen3-1.7B (47% larger).
lfm2:2.6b: Highest-quality dense model, outperforms 3B+ class models.
lfm2:2.6b-exp: Improved fine-tune of 2.6B for instruction-following, knowledge, and math.
Mixture-of-Experts (On-device MoE):
Ideal for:
Fast on-device inference (CPU/NPU optimized)
Mobile AI applications and robotics
Low-latency chat and reasoning
English/Japanese-focused tasks with strong multilingual support