TubeReads

GET IN EARLY! I'm Investing In This Breakthrough AI Chip

After attending GTC and meeting with Nvidia executives — including Jensen Huang himself — one investor is convinced Wall Street is fundamentally misunderstanding Nvidia's new Vera Rubin platform. Most analysts see just faster chips; he sees a blueprint for the entire AI revolution, with huge implications for data center spending and autonomous agents. The mainstream narrative focuses on GPU speed bumps, but the real story involves seven new chips, a radical shift from training to inference, and a $20 billion acquisition integrated in record time. Will Nvidia's orchestration of hardware, software, and the physical AI ecosystem propel it to become the world's first $10 trillion company?

Durée de la vidéo : 18:43·Publié 25 mars 2026·Langue de la vidéo : en-US
6–7 min de lecture·3,171 mots prononcésrésumé en 1,317 mots (2x)·

1

Points clés

1

Vera Rubin is a fundamentally different system from Blackwell, designed to produce as many useful tokens as possible per rack, per watt, and per dollar — optimized for OpenClaw-style agents that run for millions of tokens, not short chat prompts.

2

Nvidia's $20 billion Gro acquisition was integrated in under a year, replacing their own planned Ruben CPX chip and delivering up to 35x higher inference throughput per watt via onchip SRAM, making it one of Nvidia's most important acquisitions since Mellanox.

3

Token demand will explode as data centers serve not just billions of people, but tens of billions of always-on AI agents that call tools, browse websites, and write code — a dynamic most analysts underestimate.

4

Physical AI is already here: humanoid robots are running real warehouse shifts, and Nvidia-powered autonomous vehicles are ready to roll out on Uber's network in 2026, expanding to 28 cities by 2028.

5

Investors should watch Nvidia's data center revenue mix for signals about which workloads — training, inference, memory, or LPUs — are ramping fastest, revealing demand patterns long before they show up in headline earnings.

En bref

Nvidia's Vera Rubin platform isn't just an upgrade — it's a systemic redesign for the age of autonomous AI agents that demand millions of tokens at a time, positioning Nvidia to capture revenue across training, inference, robotics, and autonomous vehicles in ways Wall Street isn't yet pricing in.


2

Vera Rubin: A Blueprint, Not Just a Speed Bump

Rubin rewrites networking, memory, and compute for autonomous agents, not just training.

Nvidia's Vera Rubin platform represents a fundamental departure from Blackwell, not merely a generational upgrade. AI workloads are shifting from short human-written prompts to autonomous agents like OpenClaw that call tools, browse websites, write code, and run for millions of tokens at a time — workloads that cost thousands of times more tokens than regular chat. Power-efficient, low-latency inference has become the new main cost driver for AI, which is why Rubin is engineered to produce as many useful tokens as possible per rack, per watt, and per dollar.

This architectural shift has huge implications for data center spending. While most analysts predict a slowdown, the host expects data center capital expenditure to accelerate because models are now continuously fine-tuned via reinforcement learning and agents demand orders of magnitude more compute. Rubin's design — seven new chips working in concert — is purpose-built to make these OpenClaw-style agents affordable to deploy at scale, setting the stage for a new wave of infrastructure investment that Wall Street isn't yet pricing in.


3

Seven Chips, Two That Really Matter

🚀
Ruben GPU
The main AI chip with a new transformer engine delivering 5x higher inference performance, 3.5x training boost, and over 90% lower token costs versus Blackwell.
🧠
Vera CPU (88 ARM cores)
Handles orchestration, scheduling, branching logic, and confidential computing — essential for multi-agent workloads and API calls that GPUs can't do efficiently.
Gro 3 LPU
A language processing unit with 500 MB of onchip SRAM, replacing Ruben CPX and delivering 35x higher inference throughput per watt for low-latency token generation.
🔗
Bluefield 4 DPU
Data processing unit that ties GPUs, LPUs, and context memory together, enabling long-term agent context on separate drives with 5x better power efficiency.

4

The $20 Billion Gro Sprint

Nvidia integrated Gro's LPU architecture in nine months, replacing its own inference chip.

💡

The $20 Billion Gro Sprint

Nvidia announced a $20 billion deal to license Gro's technology on December 24, 2025, demoed the first Gro 3 LPX at GTC three months later, and shipped production chips within nine months. The Gro 3 LPU, built around 500 MB of onchip SRAM instead of external DRAM, quietly replaced Nvidia's own Ruben CPX accelerator and now delivers up to 35x higher inference throughput per watt and 10x more revenue per rack. The host believes this will be looked back on as Nvidia's most important acquisition since Mellanox.


5

Key Performance Gains

Rubin and Gro deliver dramatic improvements in tokens per watt and per rack.

Inference Performance Boost (Rubin GPU)
5x vs Blackwell
Driven by a new transformer engine optimized for token throughput.
Training Performance Boost (Rubin GPU)
3.5x vs Blackwell
Supports continuous fine-tuning via reinforcement learning.
Token Cost Reduction (Rubin GPU)
Over 90%
Makes large-scale agent deployment economically viable.
Inference Throughput per Watt (Gro 3 LPU)
35x higher
Compared to previous-generation inference chips, enabled by 500 MB onchip SRAM.
Tokens per Second (with STX context memory)
5x increase
Also delivers 5x better power efficiency for long-context agent workloads.
NVLink 6 Bandwidth
3.6 TB/s
Fast enough to move 250 full-length 4K movies between chips every second.

6

OpenClaw and NemoClaw: The Software Layer

OpenClaw drives token demand; NemoClaw makes agents safe for enterprises.

OPENCLAW
The Operating System for Personal AI
OpenClaw is an open-source agent that can browse the internet, code, call tools, and run for millions of tokens at a time. Jensen called it the operating system for personal AI. Data centers won't just serve billions of people, but potentially tens of billions of always-on AI agents, burning tokens to do everything humans do — except much faster and for much longer, including spinning up even more agents of their own.
NEMOCLAW
Enterprise Guard Rails and Policy Engine
OpenClaw with root access is a security nightmare for enterprises. NemoClaw is Nvidia's open-source stack that wraps OpenClaw with a policy engine, privacy routing, and secure runtime, allowing companies to decide which tools the agent can use, what data it can touch, and where everything runs. This control layer makes agents safe and deployable in the real world, unlocking enterprise adoption.

7

Physical AI: Robots and Self-Driving Cars Are Already Here

Humanoids run warehouse shifts; Nvidia autonomous vehicles will roll out on Uber in 2026.

At GTC, the host saw an entire ecosystem of robots — for warehouses, hospitals, and retail — all being trained on the same Isaac and Cosmos world model stack. Agility's Digit humanoid is already running real shifts in GXO warehouses under a robotics-as-a-service model, handling logistics for brands like Nike, Amazon, and Apple. What most investors are missing is how fast this can ramp once even a handful of designs prove themselves in the field, because a capability learned in simulation for one warehouse can be tweaked and reused for the next 100 customers.

On the autonomous vehicle side, the host spent an hour in Nvidia's L2++ Mercedes driving through downtown San Francisco, navigating double-parked cars, construction zones, and erratic drivers with ease. Nvidia-powered robo-taxis using Drive Hyperion and Alpio are planned to roll out on Uber's network in cities like LA and San Francisco as soon as next year, expanding to 28 cities through 2028. Companies like BYD, Geely, Nissan, and Isuzu are developing their own level four vehicles for ride hailing and commercial fleets. When asked about the biggest near-term application for agentic systems, Jensen said autonomous vehicles — and noted that automotive is less than 1% of Nvidia's revenue today, just as CUDA once was.


8

Why Nvidia Will Hit $10 Trillion First

Nvidia is wiring itself into every layer of the AI economy, from tokens to robots.

💡

Why Nvidia Will Hit $10 Trillion First

Nvidia isn't just selling faster GPUs. It's orchestrating tokens, agents, robots, self-driving cars, and the data centers powering it all. The company has new ways to scale beyond selling more GPU racks — layering on high-value components like Gro LPUs, Bluefield DPUs, and context memory across specialized racks. If Nvidia breaks out revenue from these components like it did for networking, the mix will tell investors which workloads are ramping fastest, revealing demand signals long before they show up in headline earnings. This is the bigger picture Wall Street is missing.


9

Titres mentionnés

NVDANvidia Corporation
UBERUber Technologies, Inc.
COHCoherent Corp.
LITELumentum Holdings Inc.

10

Personnes

Alex
Content Creator / Investor
host
Jensen Huang
CEO of Nvidia
mentioned
Spencer Huang
Nvidia Executive (Physical AI)
mentioned

Glossaire
LPU (Language Processing Unit)A specialized chip optimized for token generation in AI inference, distinct from a GPU, using onchip SRAM to store model weights and activations for ultra-low latency.
DPU (Data Processing Unit)A processor that handles networking, memory access, and data control tasks, freeing GPUs and LPUs to focus on compute-intensive AI workloads.
SRAM (Static Random Access Memory)Small, fast memory that lives on a chip, offering predictable low-latency access but at higher cost and power per bit than DRAM.
KV CacheKey-value cache used in transformer models to store intermediate attention states, reducing redundant computation during token generation.
Prefill and DecodePrefill processes the input prompt in one pass; decode generates output tokens one at a time, each requiring a full model forward pass.

Avertissement : Ceci est un résumé généré par IA d'une vidéo YouTube à des fins éducatives et de référence. Il ne constitue pas un conseil en investissement, financier ou juridique. Vérifiez toujours les informations auprès des sources originales avant de prendre des décisions. TubeReads n'est pas affilié au créateur de contenu.