We’ve redefined what’s possible with our proprietary architecture designed for efficiency, speed, and real-world deployment on any device. From wearables, to robotics, phones, laptops, cars, and more, LFMs run seamlessly on GPUs, CPUs or NPUs, making intelligence accessible everywhere.

Purpose-built for speed, capability and efficiency. Our multimodal, hybrid models run anywhere you need them and shine at agentic tasks, instruction following, data extraction, and RAG.
LFMs deliver powerful performance in a lightweight, customizable, and compute-efficient footprint for deployment in any environment.
LFM2-350M
LFM2 -700M
LFM2 -700M
LFM2-8B-A1B
Multimodal models using vision and text inputs and outputs with capabilities designed for low latency and device aware deployment.
LFM2-VL-450M
LFM2-VL-1.6B
LFM2-VL-3B
End-to-end foundation model for audio and text generation. Designed for low latency, it enables responsive, high-quality conversations with only 1.5 billion parameters.
LFM2-Audio-1.5B
Tiny customized models for specific tasks and knowledge.
Extract
Tool
Math
ColBERT
RAG
Japanese PII Extract
LFMs are rapidly customizable to deliver powerful performance for your unique use cases, devices and data. Talk to sales to see how Liquid can build solutions for you, or leverage our self-service tools to customize and deploy.
Try LEAP: Our developer-first platform creates a single workflow for customization and deployment across any operating system.
View Docs: Get started building and customizing LFMs with cookbooks, tutorials and more.
Download Models: Browse, download and build with our collections of models from Hugging Face.
Learn more about how we designed and trained our LFM2s including our hardware-in-the-loop architecture design, pre-training, knowledge distillation, and post training recipe.
Whether deploying on smartphones, laptops, vehicles, or any other device, LFMs run efficiently on CPU, GPU, and NPU hardware. Designed for millisecond latency, on-device resilience, and data privacy, LFMs unlock the full potential of local, cloud and hybrid AI across industries.

.png)
LFMs are designed for rapid customization to achieve peak performance for specified use cases at a footprint small enough to run locally on your chosen hardware.
Our full-stack solution includes architecture, optimization and deployment engines to accelerate the path from prototype to product.


Yes. Get in touch with our team to license or purchase LFMs from our library of best-in-class models.
LFMs also come with two software stacks for deployment and customization: 1) LFM inference stack and 2) LFM customization stack. We currently prioritize working with clients on enabling edge and on-prem use cases. Connect with our team to learn more about our business model.
Yes. We have built an on-prem LFM customization stack available for purchase to enterprises. LFMs can be rapidly fine-tuned, specialized, and optimized for local, private, safety-critical, and latency-bound use cases for enterprise applications – all within the security of your enterprise firewall.