News from the AI & ML world
Megan Crouse@techrepublic.com
//
Microsoft has unveiled BitNet b1.58, a groundbreaking language model designed for ultra-efficient operation. Unlike traditional language models that rely on 16- or 32-bit floating-point numbers, BitNet utilizes a mere 1.58 bits per weight. This innovative approach significantly reduces memory requirements and energy consumption, enabling the deployment of powerful AI on devices with limited resources. The model is based on the standard transformer architecture, but incorporates modifications aimed at efficiency, such as BitLinear layers and 8-bit activation functions.
The BitNet b1.58 2B4T model contains two billion parameters and was trained on a massive dataset of four trillion tokens, roughly equivalent to the contents of 33 million books. Despite its reduced precision, BitNet reportedly performs comparably to models that are two to three times larger. In benchmark tests, it outperformed other compact models and performed competitively with significantly larger and less efficient systems. Its memory footprint is just 400MB, making it suitable for deployment on laptops or in cloud environments.
Microsoft has released dedicated inference tools for both GPU and CPU execution, including a lightweight C++ version, to facilitate adoption. The model is available on Hugging Face. Future development plans include expanding the model to support longer texts, additional languages, and multimodal inputs such as images. Microsoft is also working on another efficient model family under the Phi series. The company demonstrated that this model can run on a Apple M2 chip.
ImgSrc: assets.techrepu
References :
- www.techrepublic.com: Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware
- medium.com: Microsoft has released a new language model, BitNet, designed for energy efficiency, minimizing the computational and memory requirements for use on older hardware. This strategy aims to make advanced AI more accessible to a wider range of users.
- THE DECODER: Microsoft's new model, BitNet b1.58 2B4T, is intended to operate with reduced memory and energy consumption. The model demonstrates an effort to expand access and reduce computational burdens for AI applications.
- www.zdnet.com: Microsoft introduces BitNet b1.58 2B4T, a new small language model designed to run efficiently on older hardware without GPUs.
- the-decoder.com: Microsoft Shows How to Put AI Models on a Diet
- arstechnica.com: This article details Microsoft researchers creating a super-efficient AI that uses up to 96% less energy.
- www.tomshardware.com: Microsoft researchers build 1-bit AI LLM, model small enough to run on some CPUs
- TechSpot: Microsoft's BitNet shows what AI can do with just 400MB and no GPU
- www.sciencedaily.com: Researchers developed a more efficient way to control the outputs of a large language model, guiding it to generate text that adheres to a certain structure, like a programming language, and remains error free.
Classification:
- HashTags: #BitNet #MicrosoftAI #LowEnergyAI #AI
- Company: Microsoft
- Target: AI Researchers, Enterprises
- Product: BitNet b1.58
- Feature: Energy Efficiency
- Type: AI
- Severity: Informative