Skip to main content

Nvidia GTC 2026 keynote live blog — Vera Rubin GPUs and CPUs, DLSS 5, and the 'future of technology'

A lot of AI announcements, and maybe a thing or two for consumers.

GTC 2026
(Image credit: © 3DTested)

Nvidia CEO Jensen Huang is on the stage at the SAP Center in San Jose, California to deliver the GTC 2026 keynote. 3DTested Editor-in-Chief Paul Alcorn and Senior GPU Analyst Jeffrey Kampman are on the ground in San Jose, and we'll be delivering you updates from the keynote floor in real-time. The keynote begins at 11am PT, and it's expected to run two hours.

As with any Nvidia event these days, we expect to hear a lot about AI advancements, as Nvidia continues to run circles around the tech industry in the enterprise space. We could see some announcements on the consumer front, as well, but if you're expecting to see new GeForce graphics cards, you might be leave disappointed.

Refresh

This is 'the most important chart' for companies, says Nvidia

Nvidia CEO showing a chart at GTC 2026.

(Image credit: 3DTested)

Tokens are "the new commodity," according to Nvidia. For businesses, Nvidia says that the throughput of an AI factory at iso power is something that will be "studied for years." More tokens means smarter models, and the smarter the models get, you need better token throughput. Nvidia says that, at every tier, Vera Rubin delivers much higher throughput.

Nvidia GTC 2026

(Image credit: 3DTested)

Jensen explains how NVLink for Rubin Ultra works, with compute sitting in the front and the scale-up fabric in the back.

Groq 3 LPU and Groq LPX join the fray

Nvidia CEO showing off Vera Rubin at GTC 2026.

(Image credit: 3DTested)

A new addition to the system is a Groq LPX rack, which we learned about ahead of GTC. You can read Jeffrey Kampman's breakdown of Groq 3 in Vera Rubin now.

Learn more about Nvidia's Vera CPU

Vera Rubin joins Jensen on stage

Vera Rubin at GTC 2026.

(Image credit: 3DTested)

Vera Rubin NVL72 is the "engine supercharging the era of agentic AI." A new addition is the Groq 3 LPX tray, and as a whole, Nvidia says it's delivered 40 million times more compute over the past decade. Jensen is showing off Vera Rubin on stage; the whole thing. It's "one giant system."

Jensen says the Vera CPU is designed for high single-threaded performance, and the company built it go along with its racks for agentic processing.

'It's now a factory to generate tokens'

Nvidia CEO Jensen Huang presenting at GTC 2026.

(Image credit: 3DTested)

Jensen says data centers used to be a place to store files, and they're now a factory to generate tokens. Inference is the workload and tokens are the new commodity, says Nvidia. Now, onto a short video showing how we got here.

50x performance per watt, 35x lower cost

Nvidia says Grace Blackwell was 'a giant bet'

Nvidia CEO presenting at GTC 2026.

(Image credit: 3DTested)

Nvidia NVL72 was a "giant bet," says Jensen, and he thanked Nvidia's partners for sticking with the company. "It wasn't easy for anybody... Inference is the ultimate hard." The bet paid off, according to Nvidia, which you can see in the slide above.

Nvidia is the only company that runs every domain of AI across every domain of AI models

Nvidia says it's going to double demand through the next year

GTC 2026

(Image credit: 3DTested)

Last year, Nvidia said it saw about $500 billion of high confidence demand and purchase orders for Blackwell and Rubin through 2026. "I see through 2027 at least $1 trillion," says Jensen. "Now, does it make any sense?" Jensen says that's what he's going to spend the rest of the keynote talking about.

An accelerated timeline of AI

GTC 2026

(Image credit: 3DTested)

There's been rapid AI development over the past few years. In 2023, it was ChatGPT. In 2024, it was reasoning models like o1, and in 2025, it was huge models with massive context windows like Claude Code. It's the first "agentic model," says Jensen. The executive says 100% of Nvidia is using Claude Code, along with other models. In 2026, Nvidia says we've reached an "inflection point for inference."

Nvidia 'reinvented computing'

cuDNN is what caused the 'big bang' of AI

A video demo at GTC 2026.

(Image credit: 3DTested)

Nvidia says cuDNN, or CUDA Deep Neural Network, is one of the most important libraries the company has ever made, saying it caused the "big bang" of modern AI. Nvidia is showing a short video about its various CUDA-X libraries, including a life-like video that's entirely simulated.

Bringing it back to CUDA

Nvidia says it needs domain-specific libraries to address the needs of different industries

GTC 2026

(Image credit: 3DTested)

AI has a lot of applications, but Jensen says it isn't as simple as throwing GenAI at the wall and hoping it sticks. "We have to have domain-specific libraries that solve problems in every one of these verticals," he says.

'Vertically integrated but horizontally open'

GTC 2026

(Image credit: 3DTested)

Jensen describes Nvidia as "vertically integrated but horizontally open," which may or may not raise some eyebrows at the FTC. Regardless, Nvidia says there's "no other way" it can be given what it's trying to do with accelerated computing, delivering the entire stack to customers.

Nvidia is bringing OpenAI to AWS this year

GTC 2026

(Image credit: 3DTested)

"As you know, [OpenAI] is completely compute-constrained." Jensen says that OpenAI will come to AWS this year, hopefully lightening the load on its massive infrastructure demand.

More Moore's Law talk

Nvidia CEO talking about Google Cloud.

(Image credit: 3DTested)

Jensen likes to talk about the death of Moore's Law, and he's doing so once again. "Moore's Law has run out of steam, accelerated computing allows us to take giant leaps forward." Jensen is showing off an example with Google Cloud and showing how Nvidia's acceleration can be repeated across companies and industries.

AI can solve unstructured data, says Jensen

'This is my best slide'

Nvidia CEO presenting a slide on structured data.

(Image credit: Future)

Jensen jokes that he's going to spend the rest of the keynote going through the slide you can see above about structured data. This is "the ground truth" of enterprise computing.

What is DLSS 5?

Nvidia presenting DLSS 5.

(Image credit: 3DTested)

Nvidia says it combined controllable 3D graphics and structured data with generative worlds. "This concept of fusing structured data with generative AI will repeat itself in one industry after another industry after another industry."

Nvidia is showing off the next generation of computer graphics: DLSS 5

'GeForce is Nvidia's greatest marketing campaign'

Nvidia CEO at GTC 2026.

(Image credit: 3DTested)

Jensen says that "GeForce is Nvidia's greatest marketing campaign." It's an interesting way to frame the conversation, and one that Nvidia has been trying to crack for the past few years. Jensen paints a picture of Nvidia creating the first programmable shader 25 years ago, which eventually led to CUDA, and used GeForce as a vehicle to drive adoption.

Pricing of Ampere in the cloud is going up

'We've been working on CUDA for 20 years'

Nvidia CEO at GTC 2026.

(Image credit: 3DTested)

CUDA is one of the major reasons Nvidia is in the position it's in today, and this GTC marks the 20th anniversary of CUDA. "The single hardest thing is to have built up our install base, we're in every cloud and computer company in every single industry," says Jensen.

The man of the hour is here: CEO Jensen Huang

Nvidia CEO Jensen Huang on stage at GTC 2026.

(Image credit: 3DTested)

Jensen Huang has taken the stage in a familiar leather jacket. Sorry folks, there's no special jacket this time around. Jensen is starting off the show thanking some of the people that hosted the preshow leading up to the keynote.

And we're off!

That's... A lot of country music?

Running a bit behind schedule

NVIDIA GTC Keynote 2026 - YouTube NVIDIA GTC Keynote 2026 - YouTube
Watch On

We're a few minutes past the top of the hour, and we're still waiting on the keynote to start. In the meantime, a quick reminder that you can watch along with us through the live stream above.

T-Minus 5 Minutes

The GTC 2026 keynote stage.

(Image credit: 3DTested.)

Run to the bathroom, get your drink ready, and settle in. We're just a few minutes away from the start of GTC 2026. Jensen will probably start with a short history of Nvidia's role in AI, but we expect the announcements to rapid-fire out after that point. We're sat down in the SAP Center in San Jose and ready to dig in.

What to expect from GTC 2026

  • Intel x Nvidia partnership — Nvidia bought $5 billion in Intel stock last year, and at the time, announced that the two companies would be working together on custom x86 processors across both the data center and consumer PCs. The deal has apparently been decades in the making. It's not clear if we'll hear about consumer or enterprise chips, or both, but there's a good chance we'll hear something from the partnership.
  • The 'future of real-time rendering' Nvidia presented at GDC (not GTC) about neural rendering, but just a week later, the company is teasing that it will reveal the "future of real-time rendering" at GTC (not GDC). Maybe it's a new DLSS feature, maybe it's something completely new. We don't know, but Nvidia has already confirmed something is coming for gamers during the keynote.
  • More on Vera Rubin Nvidia officially launched its Vera Rubin NVL72 in January, and it started shipping samples to customers just weeks ago. These next-gen AI data center boards are on-track for the second half of the year, so we expect to hear a lot about them during the keynote.
  • AI agents — Since the release of OpenClaw, the tech industry has been washed in talk of AI agents. Nvidia will talk about AI agents during the keynote, that much is almost guaranteed. We could see an announcement of "NemoClaw," which is an AI agent Nvidia is reportedly developing to compete with OpenClaw.
  • Nvidia N1/N1X — Perhaps the biggest rumor around Nvidia over the past year has been the N1 and N1X, which are two SoCs reportedly being developed for the consumer market. Do we finally see a reveal at this year's GTC? Perhaps, but this is the last item on this list for a reason.

We're moments away from GTC 2026

  • S58_is_the_goat
    Hey Jensen can ai bring the price of ram down? Thanx bye...
    Reply
  • abufrejoval
    At the heart of this new direction, stands a deep crisis: where for many years the industry thought that the uncanny valley of CGI could be overcome via advances in ray tracing algorithms, Nvidia and others could clearly see that this wasn't going to happen: even at exponential effort for CGI, realism wasn't approaching reality.

    But AI diffusion models delivered awsome results! And that's why some years ago Nvidia decided to change tracks completely, while finding intermediate technologies and products to pay for the transition.

    The idea is to completely replace the algorithmic/mathematical mix of simulation models for game (or in fact any visual) content with an approach that is purely based on machine learning models.

    Perhaps still as an intermediate it might use the same inputs that most game engines deliver to GPUs, the 3D meshes and the textures, but then apply them not in the current algorithmic manner, but mostly as "prompts" for a render that's similar to diffusion and/or how they interpolate frames today.

    If you took the famous teapot as geometry data, I guess it's quite easy to imagine how a "post-Phong-render" or even post-raytrace-render AI-PU might be able to generate something that looks way more naturalistic, even if it has nothing todo with what the designer of the teapot had in mind "exactly".

    The problem might be that than an AI would really need the information that the triangle mesh is supposed to represent a teapot, to do a good job, identifying the teapot via a recogntion model using a simple internal trangle mesh render might work, but ideally you'd want to use the highest level of annoation you can get, which is from the brain of the game designer, if those are still around.

    In short the result looking natural is vastly more important in a game, than the result being authentic or an exact reproduction of the designer's intent (or even the strict adherence to the triangle mesh). And who in Hollywood was ever interested in realism (warts!), when they sold art or illusion?

    Of course modern scene data is slightly more complex than a teapot and re-interpreting a scene for which you have no game developer's abstracted story data, but Nvidia might want to sell a few generations of products and game developers would need to adapt, too.

    If you look at video generation tools today, their abilty to produce short bursts of content that looks incredibly realistic, can be quite good, it's mostly when things get longer or scenes/perspectives change completely where things go off the rails. With NPCs or map data the huge currently unexploited advantage there is that their future actions, position, appearance etc. Can be known and thus "prompted" long before they need to be rendered, so errors don't need to accumulate as they do in current video generation models, where the prompt remains static.

    It's with vast distributed multi-player games where predictability can fail, because the shared decisions can't be agreed until real-time has passed, where I see a new wave of jitters and bugs befalling their outputs. PvE could become really rather incredible, eventually, if the money keeps flowing.

    So the biggest benefit may actually be for the content creation industry, where the last human actors may go the way of the dodo or go back to putting out a begging bowl during live performances somewhere where humans can offer them a ball of rice in return for their antics.

    BTW: I have zero insights into what Nvidia actually does. This is pure BI (bio intelligence) hallucinations from hints dropped here and there.

    P.S.: the change of approach was hinted at years ago. But the way I imagined it to go, it would have required changing the game engines to offer the high quality/high abstraction and temporally sorted metadata to the AI-PUs well in advance.

    I've heard zero reports on say Valve adding such functionality to a future release of their engine, so either I'm very wrong, or they managed to pull up a pretty impressive cover-up.
    Reply