Is Groq the David to Nvidia’s Goliath? Groq CEO is bringing the noise to VB Transform 2024

by | Jul 1, 2024 | Technology

The race is on, and the revolution will be televised. Chatter about Groq started on the site formerly known as Twitter when Matt Shumer, CEO of HypeWriteAI and OthersideAI, waxed poetic about Groq’s AI PU Inference Engine, reporting Mixtral serving at nearly 500 tok/s (in other words, delivering instantaneous answers), and shortly after served up an eye-watering demo. Groq CEO Jonathan Ross even visited CNN to show off the engine, blowing a reporter’s mind.

And it’s all in service of delivering a stable alternative to Nvidia’s GPU supremacy. Nvidia is betting that with its earnings up 265% from last year, it’s going to hang on to that dominance. Groq says at worst its technology uses about a third of the power of a GPU, but most workloads use as little as a tenth of the power. And in a world where it seems like those LLM workloads will never stop scaling, and energy demand will just keep growing (while fun fact, the earth is teetering on the edge of climate disaster), well, Groq is a major challenger to the GPU-dominated, incredibly bottlenecked, resource-starved compute landscape.

In fact, Ross claims that by next year, over half of the globe’s inference computing will be running on their chips. What’s that world going to look like? Ross will have the answers and a lot more at VentureBeat’s Transform 2024. He’s going to be diving into why AI inference is critical to enterprise technology, why efficiency is today’s urgent need — and why Groq is at the forefront of that revolution. He’ll be offering one of those mind-blowing live demo …

Article Attribution | Read More at Article Source

[mwai_chat context=”Let’s have a discussion about this article:nnThe race is on, and the revolution will be televised. Chatter about Groq started on the site formerly known as Twitter when Matt Shumer, CEO of HypeWriteAI and OthersideAI, waxed poetic about Groq’s AI PU Inference Engine, reporting Mixtral serving at nearly 500 tok/s (in other words, delivering instantaneous answers), and shortly after served up an eye-watering demo. Groq CEO Jonathan Ross even visited CNN to show off the engine, blowing a reporter’s mind.

And it’s all in service of delivering a stable alternative to Nvidia’s GPU supremacy. Nvidia is betting that with its earnings up 265% from last year, it’s going to hang on to that dominance. Groq says at worst its technology uses about a third of the power of a GPU, but most workloads use as little as a tenth of the power. And in a world where it seems like those LLM workloads will never stop scaling, and energy demand will just keep growing (while fun fact, the earth is teetering on the edge of climate disaster), well, Groq is a major challenger to the GPU-dominated, incredibly bottlenecked, resource-starved compute landscape.

In fact, Ross claims that by next year, over half of the globe’s inference computing will be running on their chips. What’s that world going to look like? Ross will have the answers and a lot more at VentureBeat’s Transform 2024. He’s going to be diving into why AI inference is critical to enterprise technology, why efficiency is today’s urgent need — and why Groq is at the forefront of that revolution. He’ll be offering one of those mind-blowing live demo …nnDiscussion:nn” ai_name=”RocketNews AI: ” start_sentence=”Can I tell you more about this article?” text_input_placeholder=”Type ‘Yes'”]

Share This