We’ve redefined what’s possible with our new AI architecture designed for efficiency, speed, and real-world deployment.
Purpose-built for edge AI, our hybrid models run entirely on-device and shine at agentic tasks, data extraction, RAG, creative writing, and multi-turn conversations.
Multilingual
Fine-tuning support
On-device deployment
Multilingual
Fine-tuning support
On-device deployment
Multilingual
Fine-tuning support
On-device deployment
Try LEAP, our developer-first platform for AI on the edge, proudly OS- and model-agnostic.
Whether deploying on smartphones, laptops, or vehicles, LFM2 runs efficiently on CPU, GPU, and NPU hardware. Designed for millisecond latency, on-device resilience, and data privacy, LFM2 unlocks the full potential of edge devices across industries.
Our full-stack solution includes architecture, optimization, and deployment engines to accelerate the path from prototype to product.
Download Apollo, our on-device AI playground for instant, private interactions—secure, low-latency, and cloud-free.
Yes. Get in touch with our team to license or purchase LFMs from our library of best-in-class models.
LFMs also come with two software stacks for deployment and customization: 1) LFM inference stack and 2) LFM customization stack. We currently prioritize working with clients on enabling edge and on-prem use cases. Connect with our team to learn more about our business model.
Yes. We have built an on-prem LFM customization stack available for purchase to enterprises. LFMs can be rapidly fine-tuned, specialized, and optimized for local, private, safety-critical, and latency-bound use cases for enterprise applications – all within the security of your enterprise firewall.
Read more about our research on various aspects of LFMs here , and follow us on X and LinkedIn.