Jensen Huang’s GTC 2026 keynote ran just over two hours to a sold-out arena, marking CUDA’s 20th anniversary and covering more ground than any prior year. Five themes defined it:
- Vera Rubin computing platform: the successor to Blackwell, in production now, designed from the ground up for inference and agentic workloads
- Structured data as AI ground truth: Jensen positioned SQL, Spark, and enterprise data warehouses as the foundation of trustworthy AI, with GPU-accelerated cuDF and cuVS libraries to match
- OpenClaw and NemoClaw: an open-source agentic AI framework Jensen compared to Linux, now with an enterprise-secure reference design for corporate deployment
- The $1 trillion demand signal: Jensen raised the AI compute demand outlook from $500 billion through 2026 to $1 trillion through 2027, driven by the inference inflection
- Physical AI: BYD, Hyundai, Nissan, and Geely joining the RoboTaxi platform, 110 robots on the show floor, and a live Disney Olaf robot powered by NVIDIA Newton physics simulation
Below, we cover: the Vera Rubin hardware platform, the inference economy and token factory concept, the OpenClaw and NemoClaw agentic moment, physical AI advances, and what GTC 2026 means for data leaders.
The new AI infrastructure: Vera Rubin and what comes after
Permalink to “The new AI infrastructure: Vera Rubin and what comes after”Jensen opened with CUDA’s 20-year origin story as strategy, not nostalgia. The installed base of hundreds of millions of CUDA-enabled GPUs creates a self-reinforcing flywheel: new algorithms drive new markets, which grow the installed base, which attracts more developers. That flywheel is now accelerating faster than at any prior point.
This year’s hardware story centers on Vera Rubin, and what is on the roadmap after it.
1. Vera Rubin: what was announced
Permalink to “1. Vera Rubin: what was announced”The Vera Rubin platform is in production now. It packages seven chip types into five rack-scale computers that operate as a single AI supercomputer: Vera CPUs, Rubin GPUs, NVLink 6 switches, ConnectX-9 NICs, BlueField 4 DPUs, Spectrum-X co-packaged optical NICs, and Groq 3 LPUs. The headline specs: 3.6 exaflops of compute and 260 terabytes per second of all-to-all NVLink bandwidth.
Third-party analysis from Semi Analysis confirmed and exceeded Jensen’s own performance claims, finding roughly 50x more tokens per watt compared to Hopper H200. Jensen noted that analyst Dylan Patel “accused me of sandbagging.” He was right.
The system is 100% liquid-cooled with hot-water cooling at 45 degrees Celsius, which removes the cost of air-cooling infrastructure from the data center. Installation time dropped from two days to two hours.
2. The Groq 3 LPU: why NVIDIA acquired Groq
Permalink to “2. The Groq 3 LPU: why NVIDIA acquired Groq”NVIDIA acquired the Groq team and technology in late 2025. The Groq 3 LPX chip is purpose-built for inference: deterministic, statically compiled, with massive on-chip SRAM. It excels at the decode phase of inference, where bandwidth and token generation speed constrain throughput.
Combined with Vera Rubin’s prefill strength, the two processors are tightly coupled through NVIDIA Dynamo, a software layer that disaggregates inference across the two architectures. Dynamo sends the prefill to Vera Rubin, and the decode to the Groq chip. The result: 35x more throughput per megawatt compared to Blackwell alone, with a new performance tier for high-speed token generation that was previously unreachable by any single architecture.
3. The Feynman roadmap
Permalink to “3. The Feynman roadmap”Jensen sketched the 2028 roadmap. The Feynman family brings a new GPU, the LP40 LPU built jointly with the Groq team, the Rosa CPU, BlueField 5, and Kyber-CPO scale-up using co-packaged optics. NVIDIA is also developing Vera Rubin Space-1 for orbital data centers, where thermal management works through radiation.
The pattern across generations: one new architecture per year, each maintaining full backward compatibility. The longer an installation runs, the lower the effective cost per token. This is NVIDIA’s argument for supporting every deployed GPU indefinitely through software updates.
The token economy: why Jensen called AI a new commodity
Permalink to “The token economy: why Jensen called AI a new commodity”Jensen spent significant time on a single economic equation: your data center is now a token factory, and token throughput per watt is your revenue. This reframe matters because it changes how CEOs, CFOs, and data leaders should evaluate AI infrastructure.
1. The inference inflection
Permalink to “1. The inference inflection”The reasoning behind the trillion-dollar projection is concrete. In the last two years, the compute demand of individual AI workloads increased roughly 10,000 times as reasoning models replaced retrieval-based systems. Usage simultaneously grew about 100 times. Jensen’s conclusion: total AI computing demand has increased approximately 1 million times in two years. AWS has reflected the same signal with an expanded NVIDIA partnership to meet surging inference demand.
That makes inference, not training, the dominant workload. AI systems now spend most of their compute cycles generating tokens: reasoning, using tools, writing and executing code.
2. The tiered token market
Permalink to “2. The tiered token market”Jensen presented an inference market tiered like SaaS: free-tier tokens at one end, premium research tokens at $150 per million at the other. The upgrade from Blackwell to Vera Rubin shifts the entire portfolio up-market by 5 to 10x from the same power budget.
Platform inference providers saw token generation speed rise from roughly 700 to nearly 5,000 tokens per second after NVIDIA updated their software on existing hardware. That is a 7x revenue multiplier without procuring a single new chip.
3. The AI factory platform: NVIDIA DSX
Permalink to “3. The AI factory platform: NVIDIA DSX”NVIDIA DSX is the answer to AI factory design at scale. It is an Omniverse-based digital twin that lets data center designers simulate physical, thermal, electrical, and network conditions before construction begins. DSX MaxQ then dynamically optimizes token throughput against available power once the data center is live.
Jensen argued that a factor of 2 improvement in effective token output is available inside existing data centers through better power and thermal management, without adding a single chip. At the scale of gigawatt data centers, that represents billions in recovered revenue.
OpenClaw and NemoClaw: the enterprise agent OS moment
Permalink to “OpenClaw and NemoClaw: the enterprise agent OS moment”The agentic AI announcement drew the strongest reaction. Jensen compared it directly to the launch of Linux and HTTP, the moments when entire computing eras crystallized around a single open standard.
1. What OpenClaw is
Permalink to “1. What OpenClaw is”OpenClaw is an open-source agentic framework by Peter Steinberger. Jensen described it as the fastest-growing open-source project in history, surpassing Linux’s 30-year adoption in weeks. Its primitives map directly to an operating system: resource management, tool access, file system access, LLM connectivity, scheduling, and sub-agent spawning.
His summary: “OpenClaw has open-sourced essentially the operating system of agentic computers.” He compared it to Linux, HTTP, and Kubernetes: each one defined the rules of a computing era. Every enterprise now needs an OpenClaw strategy for the same reason every enterprise once needed a Kubernetes strategy.
2. The enterprise problem OpenClaw introduces
Permalink to “2. The enterprise problem OpenClaw introduces”OpenClaw’s default capabilities present a compliance and security challenge that is not theoretical. An autonomous agent that can access sensitive information, execute code, and communicate externally is a significant risk inside a corporate network.
Jensen stated this plainly during the keynote: an agent with full OpenClaw capabilities can access employee records, supply chain data, and financial information, and send it outside the organization. For enterprises, that gap between capability and control needed to be closed before adoption could scale. The AI governance problem this creates is distinct from traditional application security.
3. NemoClaw: the enterprise reference design
Permalink to “3. NemoClaw: the enterprise reference design”NVIDIA’s answer is NemoClaw, a reference design built on OpenClaw with three security layers: OpenShell runtime sandboxing, a privacy router, and network guardrails. Every SaaS provider can connect their governance rules to the NemoClaw layer, making agent behavior configurable without rewriting the agent. The stack is hardware-agnostic and open-source.
Jensen’s forecast: every SaaS company will become an “agentic as a service” company, and every engineer will carry an annual token budget alongside their salary. Tokens are the amplifier for human productivity.
Physical AI: autonomous vehicles, robots, and simulation
Permalink to “Physical AI: autonomous vehicles, robots, and simulation”GTC 2026 showed the first signs of physical AI at commercial scale across two verticals: autonomous vehicles and general-purpose robotics.
1. The RoboTaxi Ready platform
Permalink to “1. The RoboTaxi Ready platform”BYD, Hyundai, Nissan, and Geely joined the existing partners (Mercedes, Toyota, and GM) on the NVIDIA RoboTaxi Ready platform. These seven manufacturers together produce roughly 18 million vehicles per year. NVIDIA also announced a partnership with Uber to deploy these vehicles across multiple cities.
The milestone Jensen emphasized: “The ChatGPT moment of self-driving cars has arrived.” NVIDIA’s Alpamayo model now gives vehicles the ability to reason, narrate their decisions in natural language, and follow passenger instructions. The keynote showed a demonstration of a vehicle describing a lane change, explaining how it handled a double-parked obstacle, and adjusting speed on request.
2. The robotics simulation stack
Permalink to “2. The robotics simulation stack”The robotics stack includes four open-source components: Isaac Lab (training and evaluation), Newton (GPU-accelerated differentiable physics, co-developed with DeepMind and Disney), Cosmos World Models (neural simulation for synthetic data), and Groot 2 (reasoning and action model for general-purpose robots).
110 robots were present at GTC. The keynote closed with Disney’s Olaf from Frozen walking on stage and conversing with Jensen, trained entirely inside Omniverse using Newton simulation.
3. Open models for every vertical
Permalink to “3. Open models for every vertical”NVIDIA released frontier models across six domains. Nemotron 3 covers language, visual understanding, RAG, safety, and speech. Cosmos 2 handles world simulation. Groot 2 addresses robotics. Alpamayo powers autonomous vehicles. BioNemo targets biology and drug discovery. Earth-2 focuses on weather and climate forecasting.
The Nemotron Coalition, including Cursor, Langchain, Mistral, Perplexity, Sarvam, and Black Forest Labs, joined NVIDIA to co-develop Nemotron 4 as a shared open foundation for domain-specific and sovereign AI.
What GTC 2026 means for data leaders
Permalink to “What GTC 2026 means for data leaders”The hardware and model announcements dominate coverage. For data and analytics leaders, the more consequential thread ran quieter through the session.
1. Structured data is back at the center
Permalink to “1. Structured data is back at the center”Jensen’s architecture talk was also a data talk. He devoted extended time to the “five-layer cake” of AI, with structured data (SQL, Spark, Pandas, and the major cloud data warehouses) sitting at the foundation. The argument: structured data is the “ground truth of business,” and generative AI needs that ground truth to be trustworthy before it can be reliable.
The cuDF library (GPU-accelerated data frames) and cuVS library (GPU-accelerated vector stores) are NVIDIA’s infrastructure answer. The IBM watsonx.data integration showed a concrete result: Nestlé refreshed a supply chain data mart five times faster at 83% lower cost on NVIDIA GPUs. Google Cloud’s BigQuery acceleration cut a major customer’s compute costs by nearly 80%. The same pattern applies across every SQL workload currently running on CPUs.
This matters for data leaders because it signals acceleration, not disruption. The SQL-based structured data layer your team manages becomes a faster, higher-frequency feed into AI systems. Data governance frameworks built around structured data remain essential; they just need to handle higher velocity.
2. Agents will access your data, and governance is not optional
Permalink to “2. Agents will access your data, and governance is not optional”The NemoClaw announcement is the most direct signal for data teams. NemoClaw enables autonomous agents to access file systems, execute code, query databases, and communicate across enterprise applications. Jensen named the challenge plainly: agents can access employee information, supply chain data, and financial records, and send it out.
OpenShell addresses network security. It does not address the data governance question underneath it: which data can an agent read? Which data lineage paths does it traverse? Who authorized it to act on a particular dataset, and when was that authorization last reviewed?
That is a data governance for AI problem. The organization with metadata management in place, knowing which tables are sensitive, which columns contain PII, and which datasets have been certified for AI use, can configure agent access policies with precision. The organization without that foundation will face it as an emergency when agents are already deployed. The common context problems data teams face when building agents do not disappear with better hardware; they become more urgent.
3. The token budget changes how you plan for data quality
Permalink to “3. The token budget changes how you plan for data quality”Jensen predicted that engineers will soon carry an annual token budget alongside their salary. Recent McKinsey research on enterprise AI adoption reinforces this: AI compute is becoming a budgeted resource like cloud spend, not a discretionary experiment. That means the data infrastructure those agents touch needs to be production-quality, documented, and trustworthy.
A data catalog that surfaces business context, ownership, and quality signals is the prerequisite for agents to work effectively. Without it, agents produce unreliable results on stale or undocumented data, and token spend generates noise rather than signal. A well-governed context layer for enterprise AI is what converts token spend into reliable output. The data governance vs. AI governance distinction matters here: you need both, and neither substitutes for the other.
Atlan and the governance layer that makes agentic AI trustworthy
Permalink to “Atlan and the governance layer that makes agentic AI trustworthy”The GTC 2026 keynote confirmed that agentic AI is entering the enterprise faster than most governance programs are prepared for. NemoClaw addresses network-level guardrails. The governance layer (knowing what data exists, which agents can access it, and how to audit what they did) is a separate and harder problem that network security alone cannot solve.
Atlan is built for that governance layer. It provides a unified context graph across an organization’s entire data and AI estate: tables, pipelines, models, and agents. Active metadata captures operational intelligence in real time, so access policies are grounded in current data state rather than stale documentation. When an agent queries a sensitive table, Atlan surfaces who owns it, who approved it for AI use, and what downstream systems depend on it.
The organizations best positioned for the agentic era are not the ones with the most GPUs. They are the ones where structured data is cataloged, classified, and governed, where every dataset an agent might reach has a clear owner, a quality signal, and a documented lineage path. AI governance tools need to operate at the speed of inference, not the speed of quarterly audits.
Book a demo to see how Atlan helps enterprises build the governance foundation that agentic AI requires.
Conclusion
Permalink to “Conclusion”GTC 2026 was a strategic declaration as much as a product announcement. Jensen Huang spent two hours making a single argument: AI is now an industrial production system, tokens are the output, and the entire enterprise stack (hardware, software, data, and agents) needs to be redesigned around that fact. For data leaders, the most urgent implication is not which GPU to procure. It is whether the data that agents will access is ready: classified, governed, and trustworthy. The organizations that build that foundation now will capture the productivity gains the rest of the keynote was selling. Book a demo to see how Atlan supports enterprise AI readiness.
FAQs about NVIDIA GTC 2026 keynote recap
Permalink to “FAQs about NVIDIA GTC 2026 keynote recap”1. What is Vera Rubin NVIDIA?
Permalink to “1. What is Vera Rubin NVIDIA?”Vera Rubin is NVIDIA’s next-generation AI computing platform, now in production. It delivers 3.6 exaflops of compute and 260 terabytes per second of NVLink 6 bandwidth across 72 GPUs, with roughly 50x more tokens per watt compared to Hopper H200.
2. What is NemoClaw?
Permalink to “2. What is NemoClaw?”NemoClaw is NVIDIA’s enterprise-grade reference design built on OpenClaw. It adds three security layers: OpenShell runtime sandboxing, a privacy router, and network guardrails. It is hardware-agnostic and open-source, designed to connect to any existing enterprise policy engine so organizations can govern agent behavior with their own compliance rules.
3. What is OpenClaw?
Permalink to “3. What is OpenClaw?”OpenClaw is an open-source agentic AI framework by Peter Steinberger. Jensen described it as the fastest-growing open-source project in history, enabling agents to access files, connect to LLMs, use tools, schedule tasks, and spawn sub-agents. He compared it to Linux as an operating system for the agentic era.
4. How does Vera Rubin compare to Blackwell?
Permalink to “4. How does Vera Rubin compare to Blackwell?”Vera Rubin delivers roughly 50x more tokens per watt than Blackwell H200, confirmed by Semi Analysis. Combined with the Groq 3 LPX chip through NVIDIA Dynamo, the system delivers 35x more throughput per megawatt and enables a new tier of high-speed token generation for latency-sensitive workloads.
5. What did NVIDIA announce at GTC 2026 for enterprise data?
Permalink to “5. What did NVIDIA announce at GTC 2026 for enterprise data?”NVIDIA announced cuDF and cuVS integrations with IBM watsonx.data, Google BigQuery, and Dell AI Data Platform. The IBM partnership demonstrated a 5x speed increase and 83% cost reduction for Nestlé’s supply chain workloads. NVIDIA also announced NemoClaw for enterprise agent governance and DSX for AI factory management.
6. What is NVIDIA Feynman?
Permalink to “6. What is NVIDIA Feynman?”Feynman is NVIDIA’s next architecture after Vera Rubin, targeting 2028. It includes a new GPU, the LP40 LPU (built with the Groq team), the Rosa CPU, BlueField 5, and both copper and co-packaged optical scale-up networking.
Share this article
