微軟發布強大新款AI推理晶片

微軟發布強大新款AI推理晶片

Techcrunch·

微軟發布了新款AI晶片Maia 200,專為高效、高速的AI推理而設計。這款Maia 100的後繼產品擁有超過1000億個電晶體,效能大幅提升,旨在降低AI企業的營運成本和功耗。

Image

Image

Topics

Latest

AI

Amazon

Apps

Biotech & Health

Climate

Cloud Computing

Commerce

Crypto

Enterprise

EVs

Fintech

Fundraising

Gadgets

Gaming

Google

Government & Policy

Hardware

Instagram

Layoffs

Media & Entertainment

Meta

Microsoft

Privacy

Robotics

Security

Social

Space

Startups

TikTok

Transportation

Venture

More from TechCrunch

Staff

Events

Startup Battlefield

StrictlyVC

Newsletters

Podcasts

Videos

Partner Content

TechCrunch Brand Studio

Crunchboard

Contact Us

Image

Microsoft announces powerful new chip for AI inference

Microsoft has announced the launch of its latest chip, the Maia 200, which the company describes as a silicon workhorse designed for scaling AI inference.

The 200, which follows the company’s Maia 100 released in 2023, has been technically outfitted to run powerful AI models at faster speeds and with more efficiency, the company has said. Maia comes equipped with over 100 billion transistors, delivering over 10 petaflops in 4-bit precision and approximately 5 petaflops of 8-bit performance—a substantial increase over its predecessor.

Inference refers to the computing process of running a model, in contrast with the compute required to train it. As AI companies mature, inference costs have become an increasingly important part of their overall operating cost, leading to renewed interest in ways to optimize the process.

Microsoft is hoping that the Maia 200 can be part of that optimization, making AI businesses run with less disruption and lower power use. “In practical terms, one Maia 200 node can effortlessly run today’s largest models, with plenty of headroom for even bigger models in the future,” the company said.

Microsoft’s new chip is also part of a growing trend of tech giants turning to self-designed chips as a way to lessen their dependence on NVIDIA, whose cutting-edge GPUs have become increasingly pivotal to AI companies’ success. Google, for instance, has its TPU, the tensor processing units—which aren’t sold as chips but as compute power made accessible through its cloud. Then there’s Amazon Trainium, the e-commerce giant’s own AI accelerator chip, which just launched its latest version, the Trainium3, in December. In each case, the TPUs can be used to offload some of the compute that would otherwise be assigned to NVIDIA GPUs, lessening the overall hardware cost.

With Maia, Microsoft is positioning itself to compete with those alternatives. In its press release Monday, the company noted that Maia delivers 3x the FP4 performance of third generation Amazon Trainium chips, and FP8 performance above Google’s seventh generation TPU.

Microsoft says that Maia is already hard at work fueling the company’s AI models from its Superintelligence team. It has also been supporting the operations of Copilot, its chatbot. As of Monday, the company said it has invited a variety of parties — including developers, academics, and frontier AI labs — to use its Maia 200 software development kit in their workloads.

Techcrunch

相關文章

  1. 微軟推出Maia 200 AI晶片,加速模型推理

    Hacker News · 3 個月前

  2. Maia 200:專為推論打造的人工智慧加速器

    Hacker News · 3 個月前

  3. 微軟發表Maia 200 AI晶片,聲稱效能超越亞馬遜與Google

    Hacker News · 3 個月前

  4. 微軟最新AI晶片與亞馬遜、Google正面較勁

    Hacker News · 3 個月前

  5. 微軟執行長納德拉表示,即使推出自家AI晶片,仍將持續採購輝達和AMD的產品

    3 個月前