Loading organizations...

§ Private Profile · Cambridge, MA, USA
Liquid AI is a technology company.
Liquid AI develops and deploys advanced foundation models for efficient, general-purpose artificial intelligence. Its core offering delivers high-performance AI solutions purpose-built for real-world applications across diverse scales. The company focuses on adaptable, resource-optimized AI architectures for dynamic, practical model deployments.
Co-founded by CEO Ramin Hasani and CSO Alexander Amini, Liquid AI emerged as an MIT spin-off. Both leveraged their expertise as machine learning and AI scientists from the Massachusetts Institute of Technology's Computer Science and Artificial Intelligence Lab. Their research informed the development of versatile, efficient foundation models, addressing AI scalability.
Liquid AI's technology serves enterprises integrating sophisticated AI, offering solutions for diverse applications including data analysis and complex problem-solving. Its vision is to establish efficient, general-purpose artificial intelligence deployable effectively at every scale, fostering widespread, practical adoption of advanced AI.
Liquid AI has raised $294.0M across 3 funding rounds.
Liquid AI has raised $294.0M in total across 3 funding rounds.
Liquid AI has raised $294.0M in total across 3 funding rounds.
Liquid AI's investors include Mathew Hein, Bain Capital, Foundation Capital, Founders Fund, NextView Ventures, Plug & Play Ventures, RET Ventures, Bill Ackman, Marc Baghadjian, Joseph Jacks, 7BC Venture Capital, Accel.
Liquid AI has raised $294.0M across 3 funding rounds. Most recently, it raised $250.0M Series A in December 2024.
Liquid AI is a Boston-based foundation model company spun out of MIT, specializing in high-performance, efficient AI systems optimized for compute-constrained environments like smartphones, laptops, vehicles, and embedded devices.[1][2] It builds Liquid Foundation Models (LFMs)—small language models and general-purpose models that process complex sequential and multimodal data (text, audio, video, time series, signals) with superior speed, memory efficiency, and reliability, serving enterprises, startups, and developers needing on-device AI for privacy, latency, and security-critical applications.[1][3] The company solves the problem of deploying powerful AI in real-world edge settings where traditional transformer models falter due to high compute demands, offering custom solutions via proprietary device-aware architecture search and tools like the LEAP developer platform and Apollo app.[3] Growth momentum includes rapid releases like the LFM2 family (e.g., 350M to 1.6B parameters across modalities) and a startup program providing full-stack access with engineering guidance.[3]
Liquid AI was founded in 2023 as an MIT CSAIL spin-out by four experts: Ramin Hasani, Mathias Lechner, Alexander Amini, and Daniela Rus, who brought deep backgrounds in dynamical systems, signal processing, numerical linear algebra, and robotics.[2] The idea emerged from their MIT research on "liquid neural networks"—dynamic, adaptive architectures inspired by biological systems, designed for efficiency and interpretability in resource-limited settings.[1][2] Early traction came from leveraging this first-principles approach to create white-box models that outperform transformers in edge deployment, quickly attracting enterprise partners and investors as collaborators in scaling production-ready AI.[1][2]
Liquid AI stands out through its foundational innovations and deployment focus:
Liquid AI rides the edge AI wave, capitalizing on exploding demand for on-device intelligence amid rising data privacy regulations (e.g., GDPR), 5G/6G latency needs, and hardware like NPUs in phones/cars.[1][3] Timing is ideal as transformer scaling hits compute walls—Liquid's efficient models enable AI ubiquity without cloud dependency, influencing robotics, wearables, and autonomous systems.[1] Market forces like chip shortages and energy costs favor its approach, while its MIT roots and open tools democratize high-capability AI, accelerating adoption across industries and fostering a developer community for specialized deployments.[2][3]
Liquid AI is poised to dominate efficient foundation models, with LFM3+ iterations expanding multimodal capabilities and enterprise integrations, potentially capturing edge AI market share as devices ship with built-in NPU support.[3] Trends like federated learning, regulatory pushes for transparent AI, and hybrid cloud-edge computing will propel it, evolving its influence from model provider to full-stack enabler for trustworthy, scalable intelligence.[1][2] As edge constraints tighten, Liquid's "flowing" adaptability positions it to redefine general-purpose AI, bringing advanced capabilities directly where they're needed most—efficient, private, and reliable at every scale.[1]