
Sign up to save your podcasts
Or


The October 7, 2025 technical release by Liquid AI introducing their new model, **LFM2-8B-A1B**, an **on-device Mixture-of-Experts (MoE)** designed for efficiency on consumer hardware. This model boasts **8.3 billion total parameters** but only uses **1.5 billion active parameters per token**, allowing it to achieve larger model quality with significantly reduced compute requirements. The document highlights the model's superior **quality and speed** compared to similar-sized dense models, detailing its architecture which is optimized for **low-latency and energy consumption** on devices like phones and laptops. Furthermore, the text presents extensive **evaluation benchmarks** across knowledge, instruction following, math, and coding tasks, demonstrating strong performance and outlining the customized **inference stacks** developed for both CPU and GPU to maximize the model’s efficiency.
Source:
https://www.liquid.ai/blog/lfm2-8b-a1b-an-efficient-on-device-mixture-of-experts
By mcgrofThe October 7, 2025 technical release by Liquid AI introducing their new model, **LFM2-8B-A1B**, an **on-device Mixture-of-Experts (MoE)** designed for efficiency on consumer hardware. This model boasts **8.3 billion total parameters** but only uses **1.5 billion active parameters per token**, allowing it to achieve larger model quality with significantly reduced compute requirements. The document highlights the model's superior **quality and speed** compared to similar-sized dense models, detailing its architecture which is optimized for **low-latency and energy consumption** on devices like phones and laptops. Furthermore, the text presents extensive **evaluation benchmarks** across knowledge, instruction following, math, and coding tasks, demonstrating strong performance and outlining the customized **inference stacks** developed for both CPU and GPU to maximize the model’s efficiency.
Source:
https://www.liquid.ai/blog/lfm2-8b-a1b-an-efficient-on-device-mixture-of-experts