📋 Model Description


library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language:
  • en
  • ar
  • zh
  • fr
  • de
  • ja
  • ko
  • es
pipeline_tag: text-generation tags:
  • liquid
  • unsloth
  • lfm2.5
  • edge
base_model:
  • LiquidAI/LFM2.5-1.2B-Instruct

[!NOTE]

Includes Unsloth chat template fixes!
For llama.cpp, use --jinja

>



Unsloth Dynamic 2.0 achieves superior accuracy & outperforms other leading quants.




src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/2b08LKpev0DNEk6DlnWkY.png"
alt="Liquid AI"
style="width: 100%; max-width: 100%; height: auto; display: inline-block; margin-bottom: 0.5em; margin-top: 0.5em;"
/>

LFM2.5-1.2B-Instruct

LFM2.5 is a new family of hybrid models designed for on-device deployment. It builds on the LFM2 architecture with extended pre-training and reinforcement learning.

  • Best-in-class performance: A 1.2B model rivaling much larger models, bringing high-quality AI to your pocket.
  • Fast edge inference: 239 tok/s decode on AMD CPU, 82 tok/s on mobile NPU. Runs under 1GB of memory with day-one support for llama.cpp, MLX, and vLLM.
  • Scaled training: Extended pre-training from 10T to 28T tokens and large-scale multi-stage reinforcement learning.

!image

Find more information about LFM2.5 in our blog post.

🗒️ Model Details

ModelParametersDescription
LFM2.5-1.2B-Base1.2BPre-trained base model for fine-tuning
LFM2.5-1.2B-Instruct1.2BGeneral-purpose instruction-tuned model
LFM2.5-1.2B-JP1.2BJapanese-optimized chat model
LFM2.5-VL-1.6B1.6BVision-language model with fast inference
LFM2.5-Audio-1.5B1.5BAudio-language model for speech and text I/O
LFM2.5-1.2B-Instruct is a general-purpose text-only model with the following features:
  • Number of parameters: 1.17B
  • Number of layers: 16 (10 double-gated LIV convolution blocks + 6 GQA blocks)
  • Training budget: 28T tokens
  • Context length: 32,768 tokens
  • Vocabulary size: 65,536
  • Languages: English, Arabic, Chinese, French, German, Japanese, Korean, Spanish
  • Generation parameters:
- temperature: 0.1 - top_k: 50 - top_p: 0.1 - repetition_penalty: 1.05
ModelDescription
LFM2.5-1.2B-InstructOriginal model checkpoint in native format. Best for fine-tuning or inference with Transformers and vLLM.
LFM2.5-1.2B-Instruct-GGUFQuantized format for llama.cpp and compatible tools. Optimized for CPU inference and local deployment with reduced memory usage.
LFM2.5-1.2B-Instruct-ONNXONNX Runtime format for cross-platform deployment. Enables hardware-accelerated inference across diverse environments (cloud, edge, mobile).
We recommend using it for agentic tasks, data extraction, and RAG. It is not recommended for knowledge-intensive tasks and programming.

Chat Template

LFM2.5 uses a ChatML-like format. See the Chat Template documentation for details. Example:

<|startoftext|><|im_start|>system
You are a helpful assistant trained by Liquid AI.<|im_end|>
<|im_start|>user
What is C. elegans?<|im_end|>
<|im_start|>assistant

You can use tokenizer.applychattemplate() to format your messages automatically.

Tool Use

LFM2.5 supports function calling as follows:

  1. Function definition: We recommend providing the list of tools as a JSON object in the system prompt. You can also use the tokenizer.applychattemplate() function with tools.
  2. Function call: By default, LFM2.5 writes Pythonic function calls (a Python list between <|toolcallstart|> and <|toolcallend|> special tokens), as the assistant answer. You can override this behavior by asking the model to output JSON function calls in the system prompt.
  3. Function execution: The function call is executed, and the result is returned as a "tool" role.
  4. Final answer: LFM2 interprets the outcome of the function call to address the original user prompt in plain text.

See the Tool Use documentation for the full guide. Example:

<|startoftext|><|im_start|>system
List of tools: [{"name": "getcandidatestatus", "description": "Retrieves the current status of a candidate in the recruitment process", "parameters": {"type": "object", "properties": {"candidateid": {"type": "string", "description": "Unique identifier for the candidate"}}, "required": ["candidateid"]}}]<|im_end|>
<|im_start|>user
What is the current status of candidate ID 12345?<|im_end|>
<|im_start|>assistant
<|toolcallstart|>[getcandidatestatus(candidateid="12345")]<|toolcallend|>Checking the current status of candidate ID 12345.<|imend|>
<|im_start|>tool
[{"candidateid": "12345", "status": "Interview Scheduled", "position": "Clinical Research Associate", "date": "2023-11-20"}]<|imend|>
<|im_start|>assistant
The candidate with ID 12345 is currently in the "Interview Scheduled" stage for the position of Clinical Research Associate, with an interview date set for 2023-11-20.<|im_end|>

🏃 Inference

LFM2.5 is supported by many inference frameworks. See the Inference documentation for the full list.

NameDescriptionDocsNotebook
TransformersSimple inference with direct access to model internals.LinkColab link
vLLMHigh-throughput production deployments with GPU.LinkColab link
llama.cppCross-platform inference with CPU offloading.LinkColab link
MLXApple's machine learning framework optimized for Apple Silicon.Link
LM StudioDesktop application for running LLMs locally.Link
Here's a quick start example with Transformers:
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer

model_id = "LiquidAI/LFM2.5-1.2B-Instruct"
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto",
dtype="bfloat16",

attnimplementation="flashattention_2" <- uncomment on compatible GPU


)
tokenizer = AutoTokenizer.frompretrained(modelid)
streamer = TextStreamer(tokenizer, skipprompt=True, skipspecial_tokens=True)

prompt = "What is C. elegans?"

inputids = tokenizer.applychat_template(
[{"role": "user", "content": prompt}],
addgenerationprompt=True,
return_tensors="pt",
tokenize=True,
).to(model.device)

output = model.generate(
input_ids,
do_sample=True,
temperature=0.1,
top_k=50,
top_p=0.1,
repetition_penalty=1.05,
maxnewtokens=512,
streamer=streamer,
)

🔧 Fine-Tuning

We recommend fine-tuning LFM2.5 for your specific use case to achieve the best results.

NameDescriptionDocsNotebook
SFT (Unsloth)Supervised Fine-Tuning with LoRA using Unsloth.LinkColab link
SFT (TRL)Supervised Fine-Tuning with LoRA using TRL.LinkColab link
DPO (TRL)Direct Preference Optimization with LoRA using TRL.LinkColab link

📊 Performance

Benchmarks

We compared LFM2.5-1.2B-Instruct with relevant sub-2B models on a diverse suite of benchmarks.

ModelGPQAMMLU-ProIFEvalIFBenchMulti-IFAIME25BFCLv3
LFM2.5-1.2B-Instruct38.8944.3586.2347.3360.9814.0049.12
Qwen3-1.7B34.8542.9173.6821.3356.489.3346.30
Granite 4.0-1B24.2433.5379.6121.0043.653.3352.43
Llama 3.2 1B Instruct16.5720.8052.3715.9330.160.3321.44
Gemma 3 1B IT24.2414.0463.2520.4744.311.0016.64
GPQA, MMLU-Pro, IFBench, and AIME25 follow ArtificialAnalysis's methodology. For IFEval and Multi-IF, we report the average score across strict and loose prompt and instruction accuracies. For BFCLv3, we report the final weighted average score with a custom Liquid handler to support our tool use template.

Inference speed

LFM2.5-1.2B-Instruct offers extremely fast inference speed on CPUs with a low memory profile compared to similar-sized models.

!image

In addition, we are partnering with AMD, Qualcomm, and Nexa AI to bring the LFM2.5 family to NPUs. These optimized models are available through our partners, enabling highly efficient on-device inference.

DeviceInferenceFrameworkModelPrefill (tok/s)Decode (tok/s)Memory (GB)
Qualcomm Snapdragon® X EliteNPUNexaMLLFM2.5-1.2B-instruct2591630.9GB
Qualcomm Snapdragon® Gen4 (ROG Phone9 Pro)NPUNexaMLLFM2.5-1.2B-instruct4391820.9GB
Qualcomm Snapdragon® Gen4 (Samsung Galaxy S25 Ultra)CPUllama.cpp (Q4_0)LFM2.5-1.2B-instruct33570719MB
Qualcomm Snapdragon® Gen4 (Samsung Galaxy S25 Ultra)CPUllama.cpp (Q4_0)Qwen3-1.7B181401306MB
These capabilities unlock new deployment scenarios across various devices, including vehicles, mobile devices, laptops, IoT devices, and embedded systems.

Contact

For enterprise solutions and edge deployment, contact [email protected].

Citation

@article{liquidai2025lfm2,
  title={LFM2 Technical Report},
  author={Liquid AI},
  journal={arXiv preprint arXiv:2511.23404},
  year={2025}
}

📂 GGUF File List

📁 Filename 📦 Size ⚡ Download
LFM2.5-1.2B-Instruct-BF16.gguf
LFS FP16
2.18 GB Download
LFM2.5-1.2B-Instruct-Q2_K.gguf
LFS Q2
461.01 MB Download
LFM2.5-1.2B-Instruct-Q2_K_L.gguf
LFS Q2
461.01 MB Download
LFM2.5-1.2B-Instruct-Q3_K_M.gguf
LFS Q3
572.54 MB Download
LFM2.5-1.2B-Instruct-Q3_K_S.gguf
LFS Q3
532.3 MB Download
LFM2.5-1.2B-Instruct-Q4_0.gguf
Recommended LFS Q4
663.52 MB Download
LFM2.5-1.2B-Instruct-Q4_1.gguf
LFS Q4
725.27 MB Download
LFM2.5-1.2B-Instruct-Q4_K_M.gguf
LFS Q4
697.04 MB Download
LFM2.5-1.2B-Instruct-Q4_K_S.gguf
LFS Q4
668.02 MB Download
LFM2.5-1.2B-Instruct-Q5_K_M.gguf
LFS Q5
804.29 MB Download
LFM2.5-1.2B-Instruct-Q5_K_S.gguf
LFS Q5
787.02 MB Download
LFM2.5-1.2B-Instruct-Q6_K.gguf
LFS Q6
918.24 MB Download
LFM2.5-1.2B-Instruct-Q8_0.gguf
LFS Q8
1.16 GB Download
LFM2.5-1.2B-Instruct-UD-Q2_K_XL.gguf
LFS Q2
461.01 MB Download
LFM2.5-1.2B-Instruct-UD-Q3_K_XL.gguf
LFS Q3
572.54 MB Download
LFM2.5-1.2B-Instruct-UD-Q4_K_XL.gguf
LFS Q4
697.04 MB Download
LFM2.5-1.2B-Instruct-UD-Q5_K_XL.gguf
LFS Q5
804.29 MB Download
LFM2.5-1.2B-Instruct-UD-Q6_K_XL.gguf
LFS Q6
949.24 MB Download
LFM2.5-1.2B-Instruct-UD-Q8_K_XL.gguf
LFS Q8
1.28 GB Download