πŸ“‹ Model Description


library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language:
  • en
  • ar
  • zh
  • fr
  • de
  • ja
  • ko
  • es
pipeline_tag: text-generation tags:
  • liquid
  • lfm2
  • edge
  • moe
  • llama.cpp
  • gguf
base_model:
  • LiquidAI/LFM2-8B-A1B



src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/2b08LKpev0DNEk6DlnWkY.png"
alt="Liquid AI"
style="width: 100%; max-width: 100%; height: auto; display: inline-block; margin-bottom: 0.5em; margin-top: 0.5em;"
/>



Try LFM β€’ Documentation β€’ LEAP

LFM2-8B-A1B-GGUF

LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.

We're releasing the weights of our first MoE based on LFM2, with 8.3B total parameters and 1.5B active parameters.

  • LFM2-8B-A1B is the best on-device MoE in terms of both quality (comparable to 3-4B dense models) and speed (faster than Qwen3-1.7B).
  • Code and knowledge capabilities are significantly improved compared to LFM2-2.6B.
  • Quantized variants fit comfortably on high-end phones, tablets, and laptops.

Find more information about LFM2-8B-A1B in our blog post.

πŸƒ How to run LFM2

Example usage with llama.cpp:

llama-cli -hf LiquidAI/LFM2-8B-A1B-GGUF

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
LFM2-8B-A1B-F16.gguf
LFS FP16
15.54 GB Download
LFM2-8B-A1B-Q4_0.gguf
Recommended LFS Q4
4.41 GB Download
LFM2-8B-A1B-Q4_K_M.gguf
LFS Q4
4.7 GB Download
LFM2-8B-A1B-Q5_K_M.gguf
LFS Q5
5.51 GB Download
LFM2-8B-A1B-Q6_K.gguf
LFS Q6
6.38 GB Download
LFM2-8B-A1B-Q8_0.gguf
LFS Q8
8.26 GB Download