
When you purchase through links on our site, we may earn an affiliate commission. Here’s how it works .
(Image credit: © Tom's Hardware) Nvidia's GTC 2026 keynote has wrapped. During the two-hour (and change) presentation, Nvidia CEO Jensen Huang delivered several announcements, from new Vera CPUs to Groq LPUs , as well as laid out a vision for AI over the next 12 months.
As with any Nvidia event these days, we heard a lot about AI, as well as learned about Vera Rubin systems and DLSS 5 . See our full live blog below.
Get notified of updates 2026-03-16T14:24:00.567Z We're moments away from GTC 2026 Hello and welcome to Tom's Hardware 's live blog for the GTC 2026 keynote. Jensen Huang is moments away from taking the stage. Myself, Jake Roach, will be tending the blog while our very own Paul Alcorn and Jeffrey Kampman are on the ground to cover all of the announcements in real-time.
There's always room for surprises, especially at Nvidia's own GTC event, but there are a few key announcements we're focused in on:
Intel x Nvidia partnership — Nvidia bought $5 billion in Intel stock last year, and at the time, announced that the two companies would be working together on custom x86 processors across both the data center and consumer PCs. The deal has apparently been decades in the making . It's not clear if we'll hear about consumer or enterprise chips, or both, but there's a good chance we'll hear something from the partnership.
The 'future of real-time rendering' — Nvidia presented at GDC ( not GTC) about neural rendering, but just a week later, the company is teasing that it will reveal the "future of real-time rendering" at GTC ( not GDC). Maybe it's a new DLSS feature, maybe it's something completely new. We don't know, but Nvidia has already confirmed something is coming for gamers during the keynote.
More on Vera Rubin — Nvidia officially launched its Vera Rubin NVL72 in January, and it started shipping samples to customers just weeks ago. These next-gen AI data center boards are on-track for the second half of the year, so we expect to hear a lot about them during the keynote.
AI agents — Since the release of OpenClaw, the tech industry has been washed in talk of AI agents. Nvidia will talk about AI agents during the keynote, that much is almost guaranteed. We could see an announcement of "NemoClaw," which is an AI agent Nvidia is reportedly developing to compete with OpenClaw.
Nvidia N1/N1X — Perhaps the biggest rumor around Nvidia over the past year has been the N1 and N1X , which are two SoCs reportedly being developed for the consumer market. Do we finally see a reveal at this year's GTC? Perhaps, but this is the last item on this list for a reason.
Run to the bathroom, get your drink ready, and settle in. We're just a few minutes away from the start of GTC 2026. Jensen will probably start with a short history of Nvidia's role in AI, but we expect the announcements to rapid-fire out after that point. We're sat down in the SAP Center in San Jose and ready to dig in.
Watch On We're a few minutes past the top of the hour, and we're still waiting on the keynote to start. In the meantime, a quick reminder that you can watch along with us through the live stream above.
We're all sitting in surprise here at Tom's Hardware at the amount of country music playing before Jensen takes the stage. We're nearly a quarter past the top of the hour at this point and still waiting for the keynote to start. There's nothing wrong with country music, but rustic Americana and enterprise AI isn't a combo I'd normally expect.
We may have started a few minutes late, but the keynote has officially begun. We begin with a short video about AI tokens and all the wonderful things (according to Nvidia) we've all done with them, from healthcare to space to construction.
Jensen Huang has taken the stage in a familiar leather jacket. Sorry folks, there's no special jacket this time around. Jensen is starting off the show thanking some of the people that hosted the preshow leading up to the keynote.
CUDA is one of the major reasons Nvidia is in the position it's in today, and this GTC marks the 20th anniversary of CUDA. "The single hardest thing is to have built up our install base, we're in every cloud and computer company in every single industry," says Jensen.
The prevalence of CUDA has accelerated what Nvidia calls a "flywheel." Nvidia attracts developers who develop on CUDA, which leads to more people adopting Nvidia hardware, and the lifecycle continues. Because of this, Jensen says the price of GPUs using the now-dated Ampere architecture has actually gone up in the cloud.
Jensen says that "GeForce is Nvidia's greatest marketing campaign." It's an interesting way to frame the conversation, and one that Nvidia has been trying to crack for the past few years. Jensen paints a picture of Nvidia creating the first programmable shader 25 years ago, which eventually led to CUDA, and used GeForce as a vehicle to drive adoption.
The first announcement is a big one: DLSS 5. Nvidia is showing it off in Resident Evil: Requiem, Hogwarts Legacy, and Starfield. We're all waiting eagerly to hear more details about what DLSS 5 includes, but the side-by-side comparisons are compelling.
Nvidia says it combined controllable 3D graphics and structured data with generative worlds. "This concept of fusing structured data with generative AI will repeat itself in one industry after another industry after another industry."
Jensen jokes that he's going to spend the rest of the keynote going through the slide you can see above about structured data. This is "the ground truth" of enterprise computing.
Jensen is describing the importance of AI in unstructured data. He says this data makes up 90% of the world's data but it's been "useless" because you can't search or query it. IBM, the inventor of SQL, is accelerating WatsonX data with the cuDF acceleration framework.
Jensen likes to talk about the death of Moore's Law, and he's doing so once again. "Moore's Law has run out of steam, accelerated computing allows us to take giant leaps forward." Jensen is showing off an example with Google Cloud and showing how Nvidia's acceleration can be repeated across companies and industries.
"As you know, [OpenAI] is completely compute-constrained." Jensen says that OpenAI will come to AWS this year, hopefully lightening the load on its massive infrastructure demand.
Jensen describes Nvidia as "vertically integrated but horizontally open," which may or may not raise some eyebrows at the FTC. Regardless, Nvidia says there's "no other way" it can be given what it's trying to do with accelerated computing, delivering the entire stack to customers.
AI has a lot of applications, but Jensen says it isn't as simple as throwing GenAI at the wall and hoping it sticks. "We have to have domain-specific libraries that solve problems in every one of these verticals," he says.
"We are an algorithm company," says Jensen, after spending 10 minutes talking about the applications of Nvidia's software stack in just about every industry. Everything comes back to Nvidia's CUDA-X libraries, and Jensen describes them as the "crown jewel" of the company.
Nvidia says cuDNN, or CUDA Deep Neural Network, is one of the most important libraries the company has ever made, saying it caused the "big bang" of modern AI. Nvidia is showing a short video about its various CUDA-X libraries, including a life-like video that's entirely simulated.
Jensen is talking about some of the many "AI native" companies, which are only possible because Nvidia "reinvented computing," says the executive. We're at the beginning of a new platform shift, he says, and it's akin to the PC revolution. This really kicked off over the past two years with ChatGPT kicking off the generative AI era.
There's been rapid AI development over the past few years. In 2023, it was ChatGPT. In 2024, it was reasoning models like o1, and in 2025, it was huge models with massive context windows like Claude Code. It's the first "agentic model," says Jensen. The executive says 100% of Nvidia is using Claude Code, along with other models. In 2026, Nvidia says we've reached an "inflection point for inference."
Last year, Nvidia said it saw about $500 billion of high confidence demand and purchase orders for Blackwell and Rubin through 2026. "I see through 2027 at least $1 trillion," says Jensen. "Now, does it make any sense?" Jensen says that's what he's going to spend the rest of the keynote talking about.
Jensen says Nvidia is the only company that runs every domain of AI across every domain of AI models. Between Nvidia, Anthropic, and Meta SL, Jensen says that represents a third of the world's AI compute. Nvidia says it's proven that "you can build with complete confidence" as an AI infrastructure company.
Nvidia NVL72 was a "giant bet," says Jensen, and he thanked Nvidia's partners for sticking with the company. "It wasn't easy for anybody… inference is the ultimate hard." The bet paid off, according to Nvidia, which you can see in the slide above.
"Nobody believed me," says Jensen. When Nvidia says it delivered 30x better performance per watt, on NVL72, it was wrong. Apparently, it delivers 50x better performance per watt. Jensen once again goes back to Moore's Law, saying it would maybe deliver 1.5x perf in a year.
Jensen says data centers used to be a place to store files, and they're now a factory to generate tokens. Inference is the workload and tokens are the new commodity, says Nvidia. Now, onto a short video showing how we got here.
Vera Rubin NVL72 is the "engine supercharging the era of agentic AI." A new addition is the Groq 3 LPX tray, and as a whole, Nvidia says it's delivered 40 million times more compute over the past decade. Jensen is showing off Vera Rubin on stage; the whole thing. It's "one giant system." Jensen says the Vera CPU is designed for high single-threaded performance, and the company built it go along with its racks for agentic processing.
The 88-core Vera CPU fits into a rack with 256 chips, each of them liquid-cooled. If you want to dig in deep on Nvidia's latest chip that aims to take on AMD and Intel, take a look at our Vera CPU deep dive .
A new addition to the system is a Groq LPX rack, which we learned about ahead of GTC. You can read Jeffrey Kampman's breakdown of Groq 3 in Vera Rubin now .
Jensen explains how NVLink for Rubin Ultra works, with compute sitting in the front and the scale-up fabric in the back.
Tokens are "the new commodity," according to Nvidia. For businesses, Nvidia says that the throughput of an AI factory at iso power is something that will be "studied for years." More tokens means smarter models, and the smarter the models get, you need better token throughput. Nvidia says that, at every tier, Vera Rubin delivers much higher throughput.
Groq is important for Nvidia because it pushes beyond the limits of NVL72. With Groq LPX, Nvidia says it's able to deliver up to 10x in revenue to companies using Vera Rubin. It helps solve the problem of delivering low latency and high throughput, which Jensen described as "enemies of each other." Nvidia combined one chip for high throughput and one for low latency, which it achieved with disaggregated inference.
Jensen admits that the Grace Blackwell sampling had some issues, but apparently Vera Rubin sampling is going smoothly. In fact, the first Vera Rubin is system is apparently already running in Microsoft's Azure Cloud.
Vera Rubin is undoubtedly Nvidia's most ambitious system to date, featuring seven chips across five rack systems. Compared to x86 and Hopper, Nvidia says Vera Rubin is able to deliver 700 million tokens per second compared to just 2 million
Jensen is teasing next-gen Feynman systems. It has a new GPU, new LPU, new CPU called Rosa, Bluefield 5, and Kyber with copper and CPO scale up. Feynman systems are on-track for 2028, so we'll hear a lot more about them throughout the year. At next year's GTC, we'll probably run back the same talking point with Feynman that we hard about with Vera Rubin this year.
Nvidia is working on a system called Vera Rubin Space-1, which will be the first data center in space. Sounds like we're in early stages, but Nvidia has "a lot of great engineers" working on it.
Nvidia is streamlining the process of setting up an AI agent with OpenClaw. Type two lines of shell commands, and you're off to the races with an AI agent. From there, Nvidia says you just need to give it a task and let the agent run its course.
Now Jensen is describing what OpenClaw is, which is a description no one in the room at GTC actually needs. For anyone who isn't aware, it's an agent that can connect to cloud systems. It can spawn other agents, do scheduling, decompose a problem, etc. Jensen says it's an operating system. "It's no different than how Windows allowed us to make personal computers."
NemoClaw is enterprise-secure, helping protect sensitive information. AI agents can communicate externally and execute without intervention, which is obviously a problem. NemoClaw provides a reference software stack for businesses to keep OpenClaw secure.
Nvidia says that Nemotron 3 Ultra will be the best base model in the world. In order to scale out Nemotron, Nvidia is creating a coalition for Nemotron 4, including companies like Black Forest Labs, Perplexity, Mistral, and Cursor.
Nvidia is showing off 110 robots at GTC, showcasing its "physical AI." Nvidia announced several new partners, including four new partners for robo-taxis, including BYD, Hyundai, and Nissian. Nvidia is also partnering with Uber, connecting robo-taxis into Uber's network in select cities.
In a not-at-all-awkward meeting, Olaf from Frozen joins Jensen on stage. The executive is now describing the various AI models used to make Olaf, which is… something. Anyway, Olaf is helping close out the keynote.
Well, we're done now, I guess. Nvidia closed out its keynote with an animation of several robots (plus Jensen) sitting around a fire singing a song about the keynote. It's a country song, and probably generated by AI? I don't really know what to say about this one. A rough-talking robot singing about tokens and open-source software wasn't on my bingo card.
S58_is_the_goat Hey Jensen can ai bring the price of ram down? Thanx bye… Reply
abufrejoval At the heart of this new direction, stands a deep crisis: where for many years the industry thought that the uncanny valley of CGI could be overcome via advances in ray tracing algorithms, Nvidia and others could clearly see that this wasn't going to happen: even at exponential effort for CGI, realism wasn't approaching reality. But AI diffusion models delivered awsome results! And that's why some years ago Nvidia decided to change tracks completely, while finding intermediate technologies and products to pay for the transition. The idea is to completely replace the algorithmic/mathematical mix of simulation models for game (or in fact any visual) content with an approach that is purely based on machine learning models. Perhaps still as an intermediate it might use the same inputs that most game engines deliver to GPUs, the 3D meshes and the textures, but then apply them not in the current algorithmic manner, but mostly as "prompts" for a render that's similar to diffusion and/or how they interpolate frames today. If you took the famous teapot as geometry data, I guess it's quite easy to imagine how a "post-Phong-render" or even post-raytrace-render AI-PU might be able to generate something that looks way more naturalistic, even if it has nothing todo with what the designer of the teapot had in mind "exactly". The problem might be that than an AI would really need the information that the triangle mesh is supposed to represent a teapot, to do a good job, identifying the teapot via a recogntion model using a simple internal trangle mesh render might work, but ideally you'd want to use the highest level of annoation you can get, which is from the brain of the game designer, if those are still around. In short the result looking natural is vastly more important in a game, than the result being authentic or an exact reproduction of the designer's intent (or even the strict adherence to the triangle mesh). And who in Hollywood was ever interested in realism (warts!), when they sold art or illusion? Of course modern scene data is slightly more complex than a teapot and re-interpreting a scene for which you have no game developer's abstracted story data, but Nvidia might want to sell a few generations of products and game developers would need to adapt, too. If you look at video generation tools today, their abilty to produce short bursts of content that looks incredibly realistic, can be quite good, it's mostly when things get longer or scenes/perspectives change completely where things go off the rails. With NPCs or map data the huge currently unexploited advantage there is that their future actions, position, appearance etc. can be known and thus "prompted" long before they need to be rendered, so errors don't need to accumulate as they do in current video generation models, where the prompt remains static. It's with vast distributed multi-player games where predictability can fail, because the shared decisions can't be agreed until real-time has passed, where I see a new wave of jitters and bugs befalling their outputs. PvE could become really rather incredible, eventually, if the money keeps flowing. So the biggest benefit may actually be for the content creation industry, where the last human actors may go the way of the dodo or go back to putting out a begging bowl during live performances somewhere where humans can offer them a ball of rice in return for their antics. BTW: I have zero insights into what Nvidia actually does. This is pure BI (bio intelligence) hallucinations from hints dropped here and there. P.S.: the change of approach was hinted at years ago. But the way I imagined it to go, it would have required changing the game engines to offer the high quality/high abstraction and temporally sorted metadata to the AI-PUs well in advance. I've heard zero reports on say Valve adding such functionality to a future release of their engine, so either I'm very wrong, or they managed to pull up a pretty impressive cover-up. Reply
SkyBill40 All that… and not a damned thing about consumer space GPUs. Gee, thanks, AI. Reply
RoLleRKoaSTeR Coming soon. You will never own anything – House, car, media – ever again. Nor the right to repair or modify equipment you "rent" Computer? – nope – it is now a thin client box that you will have to rent and if it needs GPU ability – you will have to rent that, too. Reply
abufrejoval RoLleRKoaSTeR said: Coming soon. You will never own anything – House, car, media – ever again. Nor the right to repair or modify equipment you "rent" Computer? – nope – it is now a thin client box that you will have to rent and if it needs GPU ability – you will have to rent that, too. Nonsense, without a job you can't pay rent, neither. Reply
usertests https://i.ytimg.com/vi/I4HmlBRWRjQ/hqdefault.jpg?sqp=-oaymwEXCOADEI4CSFryq4qpAwkIARUAAIhCGAE=&rs=AOn4CLCaZz5mULO6WiusgDDbbwe_Cs0qwA abufrejoval said: Nonsense, without a job you can't pay rent, neither. UBI by Olaf SkyBill40 said: All that… and not a damned thing about consumer space GPUs. Gee, thanks, AI. My hopes of RTX 5050 9GB: Dashed Reply
abufrejoval I felt there was a lot of Cerebras in what they do with Groq. Reply
Key considerations
- Investor positioning can change fast
- Volatility remains possible near catalysts
- Macro rates and liquidity can dominate flows
Reference reading
- https://www.tomshardware.com/news/live/SPONSORED_LINK_URL
- https://www.tomshardware.com/news/live/nvidia-gtc-2026-keynote-live-blog-jensen-huang#main
- https://www.tomshardware.com
- Nvidia launches DGX Station with its bleeding-edge GB300 Grace Blackwell Superchip — now available to order and will begin shipping in the coming months
- Grab 32GB of Corsair DDR5 RAM for just $111 in this epic Newegg combo with the 9850X3D — $1,020 bundle for an AMD gaming PC build includes an Asus X870E-E mothe
- Global chip supply chain under threat as US-Iran conflict enters third week — Strait of Hormuz blockade is days away from crippling Taiwan's semiconductor indus
- Micron enters high-volume production of HBM4 for Nvidia Vera Rubin – 2.3x bandwidth improvement and 20% boost in power efficiency
- Grab 32GB of Corsair DDR5 RAM for just $111 in this epic Newegg combo with the 9850X3D — $1,020 bundle for an AMD gaming PC build includes an Asus X870E-E mothe
Informational only. No financial advice. Do your own research.