16.1 C
New Delhi
Friday, November 22, 2024
HomeTechHow Nvidia built a competitive moat around AI chips

How Nvidia built a competitive moat around AI chips


Naveen Rao, a neuroscientist turned tech entrepreneur, once tried to compete with Nvidia, the world’s leading maker of chips tailored for artificial intelligence.


At a startup that was later bought by the semiconductor giant Intel, Rao worked on chips intended to replace Nvidia’s graphics processing units, which are components adapted for AI tasks like machine learning. But while Intel moved slowly, Nvidia swiftly upgraded its products with new AI features that countered what he was developing, Rao said.

After leaving Intel and leading a software startup, MosaicML, Rao used Nvidia’s chips and evaluated them against those from rivals. He found that Nvidia had differentiated itself beyond the chips by creating a large community of AI programmers who consistently invent using the company’s technology.

“Everybody builds on Nvidia first,” Rao said. “If you come out with a new piece of hardware, you’re racing to catch up.”

Over more than 10 years, Nvidia has built a nearly impregnable lead in producing chips that can perform complex AI tasks like image, facial and speech recognition, as well as generating text for chatbots like ChatGPT. The one-time industry upstart achieved that dominance by recognizing the AI trend early, tailoring its chips to those tasks and then developing key pieces of software that aid in AI development.

Jensen Huang, Nvidia’s co-founder and CEO, has since kept raising the bar. To maintain its leading position, his company has also offered customers access to specialized computers, computing services and other tools of their emerging trade. That has turned Nvidia, for all intents and purposes, into a one-stop shop for AI development.

Discover the stories of your interest


While Google, Amazon, Meta, IBM and others have also produced AI chips, Nvidia today accounts for more than 70% of AI chip sales and holds an even bigger position in training generative AI models, according to the research firm Omdia. In May, the company’s status as the most visible winner of the AI revolution became clear when it projected a 64% leap in quarterly revenue, far more than Wall Street had expected. On Wednesday, Nvidia – which has surged past $1 trillion in market capitalization to become the world’s most valuable chipmaker – is expected to confirm those record results and provide more signals about booming AI demand.

“Customers will wait 18 months to buy an Nvidia system rather than buy an available, off-the-shelf chip from either a startup or another competitor,” said Daniel Newman, an analyst at Futurum Group. “It’s incredible.”

Huang, 60, who is known for a trademark black leather jacket, talked up AI for years before becoming one of the movement’s best-known faces. He has publicly said that computing is going through its biggest shift since IBM defined how most systems and software operate 60 years ago. Now, he said, GPUs and other special-purpose chips are replacing standard microprocessors, and AI chatbots are replacing complex software coding.

“The thing that we understood is that this is a reinvention of how computing is done,” Huang said in an interview. “And we built everything from the ground up, from the processor all the way up to the end.”

Huang helped start Nvidia in 1993 to make chips that render images in video games. While standard microprocessors excel at performing complex calculations sequentially, the company’s GPUs do many simple tasks at once.

In 2006, Huang took that further. He announced software technology called CUDA that helped program the GPUs for new tasks, turning them from single-purpose chips to more general-purpose ones that could take on other jobs in fields like physics and chemical simulations.

A big breakthrough came in 2012, when researchers used GPUs to achieve humanlike accuracy in tasks such as recognizing a cat in an image – a precursor to recent developments like generating images from text prompts.

Nvidia responded by turning “every aspect of our company to advance this new field,” Jensen recently said in a commencement speech at National Taiwan University.

The effort, which the company estimated has cost more than $30 billion over a decade, made Nvidia more than a component supplier. Besides collaborating with leading scientists and startups, the company built a team that directly participates in AI activities like creating and training language models.

Advance warning about what AI practitioners need led Nvidia to develop many layers of key software beyond CUDA. Those included hundreds of prebuilt pieces of code called libraries that save labor for programmers.

In hardware, Nvidia gained a reputation for consistently delivering faster chips every couple of years. In 2017, it started tweaking GPUs to handle specific AI calculations.

That same year, Nvidia, which typically sold chips or circuit boards for other companies’ systems, also began selling complete computers to carry out AI tasks more efficiently. Some of its systems are now the size of supercomputers, which it assembles and operates using proprietary networking technology and thousands of GPUs. Such hardware may run weeks to train the latest AI models.

“This type of computing doesn’t allow for you to just build a chip and customers use it,” Huang said in the interview. “You’ve got to build the whole data center.”

Last September, Nvidia announced the production of new chips named H100, which it enhanced to handle so-called transformer operations. Such calculations turned out to be the foundation for services like ChatGPT, which have prompted what Huang calls the “iPhone moment” of generative AI.

To further extend its influence, Nvidia has also recently forged partnerships with Big Tech companies and invested in high-profile AI startups that use its chips. One was Inflection AI, which in June announced $1.3 billion in funding from Nvidia and others. The money was used to help finance the purchase of 22,000 H100 chips.

Mustafa Suleyman, Inflection’s CEO, said there was no obligation to use Nvidia’s products but competitors offered no viable alternative. “None of them come close,” he said.

Nvidia has also directed cash and scarce H100s lately to upstart cloud services such as CoreWeave, which allow companies to rent time on computers rather than buying their own. CoreWeave, which will operate Inflection’s hardware and owns more than 45,000 Nvidia chips, raised $2.3 billion in debt this month to help buy more.

Given the demand for its chips, Nvidia must decide who gets how many of them. That power makes some tech executives uneasy.

“It’s really important that hardware doesn’t become a bottleneck for AI or gatekeeper for AI,” said Clement Delangue, CEO of Hugging Face, an online repository for language models that collaborates with Nvidia and its competitors.

Some rivals said it was tough to compete with a company that sells computers, software, cloud services and trained AI models, as well as processors.

“Unlike any other chip company, they have been willing to openly compete with their customers,” said Andrew Feldman, CEO of Cerebras, a startup that develops AI chips.

But few customers are complaining, at least publicly. Even Google, which began creating competing AI chips more than a decade ago, relies on Nvidia’s GPUs for some of its work.

Demand for Google’s own chips is “tremendous,” said Amin Vahdat, a Google vice president and general manager of compute infrastructure. But, he added, “we work really closely with Nvidia.”

Nvidia doesn’t discuss prices or chip allocation policies, but industry executives and analysts said each H100 costs $15,000 to more than $40,000, depending on packaging and other factors – roughly two to three times more than the predecessor A100 chip.

Pricing “is one place where Nvidia has left a lot of room for other folks to compete,” said David Brown, a vice president at Amazon’s cloud unit, arguing that its own AI chips are a bargain compared with the Nvidia chips it also uses.

Huang said his chips’ greater performance saved customers money. “If you can reduce the time of training to half on a $5 billion data center, the savings is more than the cost of all of the chips,” he said. “We are the lowest-cost solution in the world.”

He has also started promoting a new product, Grace Hopper, which combines GPUs with internally developed microprocessors, countering chips that rivals say use much less energy for running AI services.

Still, more competition seems inevitable. One of the most promising entrants in the race is a GPU sold by Advanced Micro Devices, said Rao, whose startup was recently purchased by the data and AI company DataBricks.

“No matter how anybody wants to say it’s all done, it’s not all done,” Lisa Su, AMD’s CEO, said.



Source link

- Advertisment -

YOU MAY ALSO LIKE..

Our Archieves