“NVIDIA AI supercomputers are essentially ‘AI generation factories’ of this industrial revolution,” said its chief Jensen Huang in the latest earnings call.
Said : Ehasan Sabri
When Jensen Huang told AIM that the world we live in is built on NVIDIA GPUs, he wasn’t exaggerating. The company controls about 80% of the market for accelerators in the AI data centres operated by AWS, Google Cloud and Microsoft Azure.
Recently, it touched a $2 trillion market value and added $277 billion in stock market value on Thursday—Wall Street’s largest one-day gain in history.
The company reported a revenue of $22.1 billion, up 22% sequentially and a remarkable 265% year-on-year—well above the outlook of $20 billion—during the latest earnings call. “The world has reached the tipping point of a new computing era,” said Colette Kress, chief financial officer of NVIDIA, during the earnings call.
“Almost every single time you interact with ChatGPT, we’re inferencing. Every time you use Midjourney, we’re inferencing. Every time you see amazing Sora videos that are being generated or Runway, the videos that they’re editing, Firefly, NVIDIA is doing inferencing,” said Huang, at the recent earnings call. He said that its AI supercomputers are essentially AI generation factories of this industrial revolution.
There is NO Stopping for NVIDIA
From Huang personally delivering the first DGX-1 AI supercomputer to OpenAI in 2016, NVIDIA has come a long way. Today generative AI startups like Anthropic, Inflection, and xAI are among the examples that heavily rely on NVIDIA GPUs, specifically RTX 5000 and H100s, to keep their generative AI services running.
NVIDIA H100 and H200 GPUs by 2025 worth about $1 billion. Roughly 16,384 GPUs by mid-2024 and 32,768 GPUs by the end of 2025.
Older players, like the NSE-listed E2E Networks, have also been extending their footprint in the country, with Adani also expanding rapidly. Indian SaaS company Zoho is also utilising NVIDIA GPUs to build its own LLM to add GenAI capabilities across its suite in an attempt to reduce the reliance on hyperscalers.
What’s Next?
Last year, NVIDIA faced GPU demand challenges, but this year, it has improved the supply chain, according to Huang. “Our supply is improving, overall,” he said, adding that their supply chain is just doing an incredible job for them from wafers and packaging to memories, power regulators, transceivers, networking, and cables. There is still a shortage, but NVIDIA has ramped up H200 production.
NVIDIA plans to launch Blackwell, a new GPU range that promises improved AI compute performance compared to the current Hopper architecture, potentially reducing the need for multiple GPUs
Additionally, the company plans to build the next generation of modern data centres, what it refers to as AI factories, purpose-built to refine raw data and produce valuable intelligence. “Every car company in the future will have a factory that builds the cars—the actual goods, the atoms—and a factory that builds the AI for the cars, the electrons,” said Huang.
Huang is further targeting to build sovereign AI infrastructure worldwide. “What is being experienced here in the United States, in the West, will surely be replicated around the world, and these AI generation factories are going to be in every industry, every company, every region,” said Huang.
NVIDIA is set to host its flagship GTC conference at the San Jose Convention Center from March 18-21, 2024. Over 300,000 people are expected to attend this event (both in-person as well virtually). “I am going to tell everybody about a whole bunch of new things we’ve been working on the next generation of AI,” said Huang.