Wednesday, March 18

Nvidia’s changing its strategic approach to AI, going all in on inferencing and agents


Jensen Huang took the stage at Nvidia’s (NVDA) GTC event in San Jose, Calif., on Monday, clad in his usual leather jacket, to provide the world with an update about what the world’s most valuable company has been cooking up over the last few months.

Huang was as indefatigable as ever as he ran through his roughly two-and-a-half-hour keynote in front of some 30,000 attendees. But what’s come to be known as the Super Bowl of AI featured a noticeable shift in Nvidia’s overall AI strategy — a deeper focus on inferencing, or powering AI models, and agents.

Nvidia’s chips are traditionally known for their general-purpose use. They can train and run AI models, power robots, and serve as the backbone of self-driving cars.

And while Nvidia’s offerings are still the industry standard, upstart chip companies like Cerebras and Groq have begun designing and rolling out processors geared specifically toward running AI models, creating a potential threat to Nvidia’s formidable AI moat.

Huang and company answered that at GTC with a slew of announcements meant to prove Nvidia is the inferencing leader to beat, including the debut of its Groq 3 chip and rack system.

Nvidia CEO Jensen Huang participates in a Q&A at the company's annual GTC developers conference in San Jose, California, on March 17, 2026. (JOSH EDELSON / AFP via Getty Images)
Nvidia CEO Jensen Huang participates in a Q&A at the company’s annual GTC developers conference in San Jose, California, on March 17, 2026. (JOSH EDELSON / AFP via Getty Images) · JOSH EDELSON via Getty Images

Nvidia didn’t just go further with its inferencing capabilities, though. The company also showed off its addition to the much-hyped world of OpenClaw high-powered AI agents.

OpenClaw, which debuted as Clawd in November 2025 before being renamed Moltbot and finally OpenClaw in January, has taken off thanks to its ability to run AI agents powered by different AI models on users’ machines via apps like WhatsApp, Discord, Slack, and others.

Now, Nvidia is getting in on the buzz with its NemoClaw platform designed to improve the security and privacy of the agents.

“They are evolving in a big way, not only in inference, agentic, too,” TECHnalysis Research founder and chief analyst Bob O’Donnell told Yahoo Finance.

“The switch to OpenClaw, and now NemoClaw, to me, is even more indicative of this. It just shows how quickly they are reacting to the market.”

Nvidia’s decision to include Groq 3 as one of the seven chip platforms that make up Vera Rubin is part of its effort to stay ahead of the broader industry.

Nvidia signed a $20 billion deal with Groq in December, hiring founder Jonathan Ross, president Sunny Madra, and other members of the Groq team and giving Nvidia access to Groq’s intellectual property.

The results of the deal are Nvidia’s new Groq 3 language processing unit (LPU) and Groq 3 LPX server rack. That’s right, Nvidia now has graphics processing units (GPUs), LPUs, and central processing units (CPUs). It’s a lot of units.





Source link

Leave a Reply

Your email address will not be published. Required fields are marked *