π Model Description
library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language:
- en
- ar
- zh
- fr
- de
- ja
- ko
- es
- liquid
- lfm2
- edge
- moe
- llama.cpp
- gguf
- LiquidAI/LFM2-8B-A1B
alt="Liquid AI"
style="width: 100%; max-width: 100%; height: auto; display: inline-block; margin-bottom: 0.5em; margin-top: 0.5em;"
/>
LFM2-8B-A1B-GGUF
LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
We're releasing the weights of our first MoE based on LFM2, with 8.3B total parameters and 1.5B active parameters.
- LFM2-8B-A1B is the best on-device MoE in terms of both quality (comparable to 3-4B dense models) and speed (faster than Qwen3-1.7B).
- Code and knowledge capabilities are significantly improved compared to LFM2-2.6B.
- Quantized variants fit comfortably on high-end phones, tablets, and laptops.
Find more information about LFM2-8B-A1B in our blog post.
π How to run LFM2
Example usage with llama.cpp:
llama-cli -hf LiquidAI/LFM2-8B-A1B-GGUF
π GGUF File List
| π Filename | π¦ Size | β‘ Download |
|---|---|---|
|
LFM2-8B-A1B-F16.gguf
LFS
FP16
|
15.54 GB | Download |
|
LFM2-8B-A1B-Q4_0.gguf
Recommended
LFS
Q4
|
4.41 GB | Download |
|
LFM2-8B-A1B-Q4_K_M.gguf
LFS
Q4
|
4.7 GB | Download |
|
LFM2-8B-A1B-Q5_K_M.gguf
LFS
Q5
|
5.51 GB | Download |
|
LFM2-8B-A1B-Q6_K.gguf
LFS
Q6
|
6.38 GB | Download |
|
LFM2-8B-A1B-Q8_0.gguf
LFS
Q8
|
8.26 GB | Download |