menu
techminis

A naukri.com initiative

google-web-stories
Home

>

AR News

>

Cerebras j...
source image

VentureBeat

4w

read

435

img
dot

Image Credit: VentureBeat

Cerebras just announced 6 new AI datacenters that process 40M tokens per second — and it could be bad news for Nvidia

  • Cerebras Systems announced the expansion of six new AI data centers in North America and Europe to increase its inference capacity to over 40 million tokens per second.
  • The company aims to cater to the growing demand for fast AI inference and compete with Nvidia in this market segment.
  • Partnerships with Hugging Face and AlphaSense further strengthen Cerebras' position by providing wider access to their high-speed AI inference services.
  • Cerebras' specialized hardware is designed to run AI models 10 to 70 times faster than GPU-based solutions, appealing to customers seeking speed and efficiency.
  • The company focuses on high-speed inference applications like real-time voice and video processing, reasoning models, and coding, enhancing developer productivity.
  • Cerebras' investment in resilient infrastructure, like the tornado-resistant Oklahoma City facility, showcases its commitment to reliability and continuity of service.
  • The expansion comes at a time when the AI industry prioritizes inference capabilities for deploying generative AI in production applications efficiently.
  • With a significant presence in the United States, Cerebras also contributes to advancing domestic AI infrastructure and technological sovereignty.
  • Cerebras' strategy of focusing on high-speed inference over competing across all AI workloads positions them as a leader in a niche market segment.
  • Through cost-efficient services and faster AI model processing, Cerebras offers a compelling alternative to GPU-based solutions for various applications.

Read Full Article

like

26 Likes

For uninterrupted reading, download the app