NVIDIA GTC 2026: Keynote Announcements
There are product launches. And then there’s what NVIDIA just did at GTC 2026. Jensen Huang didn’t walk on stage to announce one thing. He announced everything: silicon, software, and about half the enterprise software industry along with it. If you blinked, you missed something significant. So let me save you the scroll and share my favorite announcements.

Before we dive into the announcements, I do want to emphasize what an amazing story teller Jensen Huang is. The build up in the keynote was, yet again, amazing. If you work in tech and you’ve never joined an in-person GTC, it should be on your bucket list. The speed of the keynote is bizar with announcement after announcement and demo after demo. One cool demo he showed, was the one on DLSS 5, where AI is used to enhance video images. Especially since it featured the Dutch National Football team:

Now, let’s dive into the announcements :)
New Hardware: Seven Chips, Full Production, No Waiting
The NVIDIA Vera Rubin platform is the headline. Seven new chips, five rack configurations, all in full production today: Rubin GPU, Vera CPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU, Spectrum-6 Ethernet switch, and a newly integrated Groq 3 LPU. Designed to operate as a single coherent pod-scale AI supercomputer.

The Vera Rubin NVL72 delivers up to 10x higher inference throughput per watt at one-tenth the cost per token compared to Blackwell, while training large MoE models with one-fourth the number of GPUs. The Vera CPU packs 88 custom Olympus cores with 1.2 TB/s of memory bandwidth: twice the bandwidth at half the power of a general-purpose CPU. A full rack holds 256 liquid-cooled CPUs sustaining 22,500+ concurrent environments. Alibaba, Meta, Oracle, and CoreWeave are already deploying it.
The Groq 3 LPX rack targets trillion-parameter, low-latency agentic workloads — up to 35x higher inference throughput per megawatt paired with Vera Rubin. The BlueField-4 STX adds a dedicated KV cache storage tier boosting inference throughput by up to 5x. And the Spectrum-6 SPX brings co-packaged optics with 5x greater optical power efficiency.
As a “one more thing”, Jensen also showed Rubin Ultra, and the switch he announced as the new NVLink. Long story short, you are now able to link 144 GPUs together. Mind blowing.

Software: The Stack Beneath the Models
NVIDIA Dynamo 1.0 is now generally available and open source: the inference operating system of the AI factory. It orchestrates GPU and memory resources across the cluster, routes requests intelligently, and boosts Blackwell inference performance by up to 7x. Already running at AWS, Azure, Google Cloud, Perplexity, Pinterest, and PayPal.
The NVIDIA Agent Toolkit bundles Nemotron open models, the AI-Q Blueprint for agentic search (tops DeepResearch Bench, cuts query costs 50%+), and the new OpenShell an open-source runtime enforcing policy-based security and privacy guardrails for autonomous agents. Adobe, SAP, Salesforce, ServiceNow, Siemens, CrowdStrike and others are already building on it.
NemoClaw brings Agent Toolkit to the OpenClaw community in a single install command: Nemotron models and OpenShell on your RTX PC, DGX Station, or DGX Spark. Always-on, private, locally controlled agents. Jensen compared OpenClaw to Mac and Windows for the personal AI era. Big claim. Watching the adoption curve, it doesn’t feel like a stretch.

More about the NemoClaw solution can be found in the other blog post I released today. Please find it here: NemoClaw and DGX: Powering the Era of Agentic Scaling
Ecosystem: Half the Industry Just Got On the Train
The Nemotron Coalition: Mistral AI, Black Forest Labs, Cursor, LangChain, Perplexity, Reflection AI, Sarvam, and Thinking Machines Lab is co-developing open frontier models on DGX Cloud. The first model is co-built with Mistral AI and underpins the upcoming Nemotron 4 family.
Adobe and NVIDIA announced a deep partnership: next-gen Firefly models on CUDA-X and NeMo, agentic creative workflows using OpenShell and Nemotron, and a cloud-native 3D digital twin solution for marketing built on Omniverse. Worth watching closely if you work in creative or marketing tech.
On the industrial side, Cadence, Dassault Systèmes, Siemens, and Synopsys are shipping NVIDIA-powered AI agents for chip design and EDA. Honda runs aerodynamic simulations 34x faster on Grace Blackwell. Siemens’ Digital Twin Composer is already used by Foxconn, HD Hyundai, PepsiCo, and KION. The Vera Rubin DSX reference design ties it all together: a co-designed framework spanning compute, power, cooling, and grid integration.

Because with $300B+ in equipment backlogs and 200+ gigawatts waiting in U.S. interconnection queues, energy is quietly the biggest bottleneck in AI infrastructure right now.
So, What Does It All Add Up To?
The inference layer has shipped. The agent runtime is open source. The training data pipeline is blueprinted. The hardware is in production. And an ecosystem of cloud providers, enterprise software platforms, and model builders is actively building on top of it all.
NVIDIA isn’t selling GPUs. It’s selling the operating model for AI factories, from the silicon to the agent runtime to the power grid keeping it all running. And not just for the extremely large enterprises or the Fortune 500, but for everyone.
If you weren’t paying close attention this week, now you are.
👉 Curious about a specific announcement? Feel free to reach out, happy to go deeper on any of these.







