Commmonn Ground

Tech & AI

NVIDIA GTC 2026: The AI Industry Just Shifted from Training to Doing

Jensen Huang just walked onto the SAP Center stage in San Jose to deliver what might be the most consequential keynote in NVIDIA's history. Not because of a single chip reveal — but because GTC 2026 is the moment the AI industry formally pivots from building models to deploying them.

Thirty thousand people from 190 countries packed downtown San Jose. The keynote was a full-stack declaration: new silicon, new software, new physics, new economics. Here's what actually matters and why.

Vera Rubin: 10x Cheaper Inference Changes Everything

The centrepiece of GTC 2026 is Vera Rubin — NVIDIA's successor to the Blackwell architecture that dominated 2024-2025. Named after the astronomer who proved the existence of dark matter, it's a six-chip platform: Rubin GPU, Vera CPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU, and Spectrum-6 Ethernet Switch.

The headline numbers are staggering:

  • 10x reduction in inference token cost compared to Blackwell
  • 4x fewer GPUs needed to train Mixture-of-Experts models
  • 50 petaflops of NVFP4 compute per GPU for inference
  • 288GB HBM4 memory with 3.0+ TB/s bandwidth per unit
  • 260 TB/s rack bandwidth — more than the entire internet

The 10x inference cost reduction is the number that rewrites business plans. Applications that were too expensive to run continuously — always-on AI agents, real-time video processing, continuous code review — become economically viable overnight. This isn't an incremental improvement. It's a step function that will cascade through cloud provider pricing over the next 18 months.

Vera Rubin is in full production. AWS, Google Cloud, Microsoft Azure, and Oracle will deploy Rubin-based instances in H2 2026. Microsoft's next-generation "Fairwater AI superfactories" will scale to hundreds of thousands of Vera Rubin Superchips. The biggest names in AI — OpenAI, Anthropic, Meta, xAI — have all committed.

NemoClaw: NVIDIA Enters the Agent Software Layer

Hardware was the appetiser. The more strategically significant announcement is NemoClaw — an open-source platform for building and deploying enterprise AI agents.

This is NVIDIA playing the same move that made CUDA a 20-year moat: give developers free, deeply integrated tools, make it the path of least resistance, and collect when workloads scale onto NVIDIA hardware. NemoClaw lets enterprises build autonomous agents that interact with files, apps, and workflows locally — no cloud dependency required.

GTC attendees could build their own always-on AI assistant at "Build-a-Claw" stations across the convention centre. Name it, define its personality, grant tool access, and deploy it on a DGX Spark or GeForce RTX laptop on the spot. It's a clever move — thousands of developers walk away with running agents on NVIDIA hardware.

Alongside NemoClaw, NVIDIA launched Nemotron 3 Super — a 120B-parameter open model with only 12B active parameters, purpose-built for agentic workloads. It has a 1-million-token context window — enough for an agent to hold an entire codebase or weeks of conversation history in memory without losing the plot.

The model is already deployed across Perplexity, Google Cloud, Oracle, AWS, CoreWeave, and dozens of inference providers. Companies like Palantir, Siemens, and Cadence are customising it for enterprise automation.

Physical AI: From Chatbots to Robots

GTC 2026 cemented "Physical AI" as NVIDIA's next trillion-dollar thesis. The conference featured robotics sessions from Tesla, Disney, Agility Robotics, KUKA, Universal Robots, and Waabi. Disney even showed AI-powered humanoid robots that self-balance using reinforcement learning trained in NVIDIA's Omniverse simulation.

The "three-computer" architecture tells the story: one computer trains the brain (DGX), one simulates the world (Omniverse), and one runs on the robot (Jetson/IGX). Specialist robots learn atomic skills — grasping, balancing, navigating — and over time combine them into composite capabilities. It mirrors how children learn: specialist first, generalist later.

Thinking Machines Lab announced a gigawatt-scale deployment of Vera Rubin systems. That's not a research lab — that's an industrial-scale commitment to physical AI infrastructure.

The Groq Factor and the Feynman Tease

Two more signals from GTC deserve attention.

First, the Groq integration. NVIDIA licensed Groq's dataflow architecture last year for a reported $20 billion. Groq's technology generates tokens at extreme speed — thousands per second — making it ideal for real-time AI agents. GTC hinted at inference products incorporating Groq technology, but concrete details remain scarce. The implication: NVIDIA is building a layered inference stack rather than treating every workload as a pure GPU problem.

Second, Jensen teased chips "the world has never seen before." The strongest candidate is Feynman — the architecture generation after Rubin, potentially built on TSMC's 1.6nm process with silicon photonics. If confirmed at future events, it would extend NVIDIA's roadmap visibility to three generations — an unprecedented signal that the company intends to outpace custom silicon from hyperscalers for years to come.

What This Actually Means

GTC 2026 isn't a GPU launch event disguised as a conference. It's a declaration that the AI industry has shifted from "training" to "doing."

The inference era is here. Models are getting cheaper to run. Agents are getting frameworks to operate autonomously. Robots are getting physics engines to learn from. And NVIDIA is positioning itself as the full-stack provider for all of it — from the atom to the application.

For developers: inference economics will cascade into cheaper API pricing by mid-2027. Plan for applications that run AI continuously, not just on-demand.

For investors: the new metric isn't FLOPS per dollar — it's tokens per megawatt. Companies that control the inference stack will compound value.

For everyone else: the AI you interact with daily is about to get significantly faster, cheaper, and more capable. The infrastructure announced today will power the products you use in 2027.

Frequently Asked Questions

What is NVIDIA Vera Rubin?

Vera Rubin is NVIDIA's next-generation AI compute platform, succeeding Blackwell. It features six new chips — including the Rubin GPU and Vera CPU — delivering up to 10x lower inference token costs and 4x fewer GPUs for training. Rubin-based products ship H2 2026 from every major cloud provider.

What is NemoClaw?

NemoClaw is NVIDIA's open-source platform for building enterprise AI agents — autonomous systems that execute multi-step tasks without constant human oversight. It's designed to deepen the CUDA ecosystem by making NVIDIA hardware the default deployment target for agentic AI workloads.

When is the NVIDIA GTC 2026 keynote?

Jensen Huang's keynote was on Monday, March 16, 2026 at 11 AM PT (2 AM March 17, Hong Kong time) at the SAP Center in San Jose. The full replay is available free on nvidia.com and YouTube.