Topics

Latest

AI

Amazon

Article image

Image Credits:Aleksander Kalka/NurPhoto / Getty Images

Apps

Biotech & Health

clime

Cloud Computing

Commerce

Crypto

Enterprise

EVs

Fintech

Fundraising

widget

Gaming

Google

Government & Policy

computer hardware

Instagram

Layoffs

Media & Entertainment

Meta

Microsoft

seclusion

Robotics

security measures

Social

blank

startup

TikTok

transit

speculation

More from TechCrunch

Events

Startup Battlefield

StrictlyVC

Podcasts

Videos

Partner Content

TechCrunch Brand Studio

Crunchboard

Contact Us

Microsoft researchers take they ’ve developed the largest - ordered series 1 - bit AI model , also love as a “ bitnet , ” to particular date . promise BitNet b1.58 2B4 T , it’sopenly availableunder an MIT licence and can run on CPUs , include Apple ’s M2 .

Bitnets are essentially constrict models plan to campaign on lightweight hardware . In stock fashion model , weights , the values that define the inner social system of a model , are oftenquantized so the models perform well on a wide range of machines . Quantizing the weight unit lowers the number of bit — the smallest units a information processing system can process — needed to present those free weight , enable example to unravel on french-fried potatoes with less remembering , faster .

Bitnets quantise weight into just three time value : -1 , 0 , and 1 . In hypothesis , that makes them far more memory- and computing - efficient than most models today .

The Microsoft research worker say that BitNet b1.58 2B4 T is the first bitnet with 2 billion parametric quantity , “ parameters ” being largely synonymous with “ weights . ” Trained on a dataset of 4 trillion tokens — equivalent to about 33 million books , by one estimation — BitNet b1.58 2B4 T exceed traditional example of similar size , the researchers claim .

BitNet b1.58 2B4 T does n’t sweep the floor with rival 2 billion - argument modeling , to be clear , but it apparently holds its own . agree to the researchers ’ examination , the theoretical account surpasses Meta ’s Llama 3.2 1B , Google ’s Gemma 3 1B , and Alibaba ’s Qwen 2.5 1.5B on benchmarks including GSM8 K ( a compendium of level - school - horizontal surface math problems ) and PIQA ( which prove forcible commonsense reasoning attainment ) .

Perhaps more imposingly , BitNet b1.58 2B4 T is speedier than other models of its size — in some cases , twice the velocity — while using a fraction of the memory .

There is a catch , however .

Join us at TechCrunch Sessions: AI

Exhibit at TechCrunch Sessions: AI

Achieving that performance requires using Microsoft ’s custom theoretical account , bitnet.cpp , which only works with certain computer hardware at the moment . abstracted from the list of supported chip are GPUs , which dominate the AI substructure landscape .

That ’s all to say that bitnets may hold promise , in particular for imagination - constrained equipment . But compatibility is — and will likely stay on — a big sticking point .