Caltech Researchers Take Another Stab at One-Bit AI Models

The Wall St Journal has this story/release about researchers from Caltech that have launched PrismML, a startup that has developed a mathematical breakthrough enabling the "radical compression" of large language models to a 1-bit architecture without sacrificing reasoning or performance. By reducing the precision of model weights to a single bit, the company’s flagship "Bonsai" models can achieve up to an eightfold increase in processing speed and an 80% reduction in energy consumption, allowing high-fidelity AI to run locally on smartphones and laptops. Backed by $16.25 million in funding from investors like Khosla Ventures and Cerberus Capital, PrismML has open-sourced its technology to shift the AI paradigm toward "intelligence density," prioritizing energy efficiency and hardware adaptability over the traditional pursuit of ever-larger data centers. Previous 1-bit efforts focused on convolutional neural networks (CNNs), whereas this work seems to apply to transformers and other models relevant to large language models (LLMs).

Other contents

Intel to invest in SambaNova. This is not a repost

Caltech Researchers Take Another Stab at One-Bit AI Models

Nvidia puts $2B into Marvell

BWR 11: GTC and OFC

BWR 11: GTC and OFC

Arm Chips Are Back, and This Time They Mean Business

Arm Chips Are Back, and This Time They Mean Business

Arm is Building Chips

Arm is Building Chips

What the Heck is a Groq?

What the Heck is a Groq?

Byrne-Wheeler Report Discusses AI Deals, Broadcom and Nvidia Earnings

Byrne-Wheeler Report Discusses AI Deals, Broadcom and Nvidia Earnings

Meta Bares MTIA Roadmap, Accelerates NPU Development

Meta Bares MTIA Roadmap, Accelerates NPU Development

s/CPX/LPX/g

s/CPX/LPX/g