Tuesday, September 16, 2025

Liquid AI releases on-device basis mannequin LFM2


Liquid AI mentioned that with its basis fashions, it hopes to realize the optimum stability between high quality, latency, and reminiscence for particular duties and {hardware} necessities. | Supply: Liquid AI

This week, Liquid AI launched LFM2, a Liquid Basis Mannequin (LFM) that the corporate mentioned units a brand new normal in high quality, pace, and reminiscence effectivity deployment.

Shifting massive generative fashions from distant clouds to lean, on‑system LLMs unlocks millisecond latency, offline resilience, and knowledge‑sovereign privateness. These are capabilities important for telephones, laptops, vehicles, robots, wearables, satellites, and different endpoints that should cause in actual time.

Liquid AI designed the mannequin to offer a quick on-device gen-AI expertise throughout the trade, unlocking an enormous variety of units for generative AI workloads. Constructed on a brand new hybrid structure, LFM2 delivers twice as quick decode and prefill efficiency as Qwen3 on CPU. It additionally considerably outperforms fashions in every measurement class, making them very best for powering environment friendly AI brokers, the corporate mentioned.

The Cambridge, Mass.-based firm mentioned these efficiency features make LFM2 the best selection for native and edge use circumstances. Past deployment advantages, its new structure and coaching infrastructure ship a 3 times enchancment in coaching effectivity over the earlier LFM era.

Liquid AI co-founder and director of MIT’s Laptop Science and Synthetic Intelligence Laboratory (CSAIL) Daniela Rus delivered a keynote on the Robotics Summit & Expo 2025, a robotics improvement occasion produced by The Robotic Report.

LFM2 fashions can be found at present on Hugging Face. Liquid AI is releasing them beneath an open license, which is predicated on Apache 2.0. The license permits customers to freely use LFM2 fashions for educational and analysis functions. Firms can even use the fashions commercially in the event that they’re smaller (beneath $10m income).

Liquid AI provides small multimodal basis fashions with a safe enterprise-grade deployment stack that turns each system into an AI system, regionally. This, it mentioned, provides it the chance to acquire an outsized share in the marketplace as enterprises pivot from cloud LLMs to cost-efficient, quick, non-public, and on‑prem intelligence.

What can LFM2 do?

Liquid AI mentioned LFM2 achieves 3 times quicker coaching in comparison with its earlier era. It additionally advantages from as much as two instances quicker decode and prefill pace on CPU in comparison with Qwen3. Moreover, the corporate claimed LFM2 outperforms similarly-sized fashions throughout a number of benchmark classes, together with information, arithmetic, instruction following, and multilingual capabilities.

LFM2 is supplied with a brand new structure. It’s a hybrid Liquid mannequin with multiplicative gates and quick convolutions. It consists of 16 blocks: 10 double-gated short-range convolution blocks and 6 blocks of grouped question consideration.

Whether or not it’s deployed on smartphones, laptops, or automobiles, LFM2 runs effectively on CPU, GPU, and NPU {hardware}. The corporate’s full-stack system consists of structure, optimization, and deployment engines to speed up the trail from prototype to product.

Liquid AI is releasing the weights of three dense checkpoints with 0.35B, 0.7B, and 1.2B parameters. Customers can strive them now on the Liquid Playground, Hugging Face, and OpenRouter.

How does LFM2 carry out towards different fashions?

Average score (MMLU, IFEval, IFBENCH, GSM8K, MMMLU) vs. model size.

Common rating (MMLU, IFEval, IFBENCH, GSM8K, MMMLU) vs. mannequin measurement. | Supply: Liquid AI

The corporate evaluated LFM2 utilizing automated benchmarks and an LLM-as-a-Choose framework to acquire a complete overview of its capabilities. It discovered that the mannequin outperforms similar-sized fashions throughout totally different analysis classes.

Liquid AI additionally evaluated LFM2 throughout seven widespread benchmarks overlaying information (5-shot MMLU, 0-shot GPQA), instruction following (IFEval, IFBench), arithmetic (0-shot GSM8K, 5-shot MGSM), and multilingualism (5-shot OpenAI MMMLU, 5-shot MGSM once more) with seven languages (Arabic, French, German, Spanish, Japanese, Korean, and Chinese language).

It discovered that LFM2-1.2B performs competitively with Qwen3-1.7B, a mannequin with a 47% larger parameter depend. LFM2-700M outperforms Gemma 3 1B IT, and its tiniest checkpoint, LFM2-350M, is aggressive with Qwen3-0.6B and Llama 3.2 1B Instruct.

How Liquid AI educated LFM2

To coach and scale-up LFM2, the corporate chosen three mannequin sizes (350M, 700M, and 1.2B parameters) concentrating on low-latency on-device language mannequin workloads. All fashions had been educated on 10T tokens drawn from a pre-training corpus comprising roughly 75% English, 20% multilingual, and 5% code knowledge sourced from the online and licensed supplies.

For the multilingual capabilities of LFM2 the corporate primarily targeted on Japanese, Arabic, Korean, Spanish, French, and German languages.

Throughout pre-training, Liquid AI leveraged its present LFM1-7B as a instructor mannequin in a information distillation framework. The corporate used the cross-entropy between LFM2’s scholar outputs and the LFM1-7B instructor outputs as the first coaching sign all through the complete 10T token coaching course of. The context size was prolonged throughout pretraining to 32k.

Put up-training began with a really large-scale Supervised Nice-Tuning (SFT) stage on a various knowledge combination to unlock generalist capabilities. For these small fashions, the corporate discovered it helpful to immediately practice on a consultant set of downstream duties, akin to RAG or operate calling. The dataset is comprised of open-source, licensed, in addition to focused artificial knowledge, the place the corporate ensures prime quality by means of a mix of quantitative pattern scoring and qualitative heuristics.

Liquid AI additional applies a customized Direct Choice Optimization algorithm with size normalization on a mix of offline knowledge and semi-online knowledge. The semi-online dataset is generated by sampling a number of completions from its mannequin, primarily based on a seed SFT dataset.

The corporate then scores all responses with LLM judges and creates desire pairs by combining the very best and lowest scored completions among the many SFT and on-policy samples. Each the offline and semi-online datasets are additional filtered primarily based on a rating threshold. Liquid AI creates a number of candidate checkpoints by various the hyperparameters and dataset mixtures. Lastly, it combines a number of its greatest checkpoints right into a closing mannequin by way of totally different mannequin merging strategies.

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles

PHP Code Snippets Powered By : XYZScripts.com