TubeReads

Four CEOs on the Future of AI: CoreWeave, Perplexity, Mistral, and IREN

The AI infrastructure gold rush is here, but scaling it requires navigating colliding forces: billion-dollar GPU contracts, shrinking token economics, specialized vertical models versus general-purpose platforms, and the race to bring compute to the source of power—whether that's West Texas wind farms or modular nuclear reactors. Four CEOs at the frontier of the AI economy open up about their distinct visions. CoreWeave's Michael Intrator reveals how a hedge fund pivoted from crypto mining into building supercomputers for hyperscalers, financing infrastructure through «boxes» that pay for themselves in 2.5 years. Aravind Srinivas of Perplexity makes the case that the orchestrator who can route intelligently across models—without favoring one—holds the real strategic advantage in an era of model specialization. Arthur Mensch of Mistral argues that open-source verticalized models, not closed general-purpose giants, will unlock proprietary data in aerospace, finance, and heavy manufacturing. And Daniel Roberts of IREN explains how the real constraint isn't GPUs anymore—it's time to compute, transmission capacity, and finding enough electricians willing to move to remote Texas sites powered by 100% renewable energy.

Duração do vídeo: 1:37:39·Publicado 23 de mar. de 2026·Idioma do vídeo: English
10–11 min de leitura·18,481 palavras faladasresumido para 2,001 palavras (9x)·

1

Pontos-chave

1

CoreWeave finances GPU infrastructure through self-amortizing «boxes»—contracts where Microsoft or other hyperscalers pay off principal, interest, and data center costs in 2.5 years of a 5-year deal, enabling rapid scale without traditional equity dilution.

2

Perplexity's competitive moat is multi-model orchestration: by remaining neutral and routing queries to the best specialist model (GPT, Claude, Gemini, Qwen, DeepSeek), it avoids the liability of betting on a single architecture and can deliver superior accuracy across tasks.

3

GPU obsolescence fears are «nonsense»—CoreWeave's average contract is 5 years, Ampere A100 prices have appreciated year-over-year, and the installed base continues to find profitable inference and experimental workloads long after frontier training moves to newer chips.

4

Mistral and enterprise customers are discovering that closed-source models can't leverage decades of proprietary IP; open models enable post-training customization, on-premises deployment, and deep integration with physical systems in manufacturing, finance, and science.

5

IREN built 4.5 GW of capacity—nearly the Bay Area's annual power consumption—by locating data centers at the source of stranded renewable energy in West Texas, where 45–50 GW of wind and solar exceeds the 12 GW transmission line to population centers, creating an arbitrage opportunity between cheap electrons and high-value compute.

Em resumo

The AI infrastructure race is no longer about who has the most GPUs—it's about who can finance them sustainably, route workloads intelligently across specialized models, and build data centers where cheap renewable power actually exists, all while training a new generation of trades workers to make it happen on the ground.


2

From Crypto Mining to AI Infrastructure: CoreWeave's Pivot

CoreWeave bootstrapped a neocloud by mining Ethereum, then donated A100s to open-source AI researchers.

Michael Intrator didn't set out to build an AI cloud—he ran an algorithmic hedge fund focused on natural gas. In 2017, he and his team got interested in cryptocurrency and began mining Ethereum with GPUs, viewing compute as an option on multiple use cases. After weathering two crypto winters, they moved up the complexity stack: CGI rendering, batch computing for medical research, and eventually neural networks. In 2020–21, CoreWeave bought A100s and donated them to Luther AI, an open-source project, reasoning that researchers couldn't complain about service-level agreements if the GPUs were free. The feedback loop was immediate: «We need more of this, you got to work on this.» When those volunteers returned to their day jobs at research labs and hyperscalers, they demanded the same infrastructure. That launched CoreWeave's commercial business.

The company understood early that computing decommoditizes at scale—anyone can run a GPU, but can you run a cluster large enough to train a world-changing model? CoreWeave focused on delivering supercomputers purpose-built for AI, living «above the Nvidia GPUs but below the models.» Their first large commercial customer was Inflection; soon they were serving OpenAI, hyperscalers, and foundation-model builders. The thesis: as scaling laws drive larger training runs, only a few players can deliver infrastructure that doesn't bottleneck innovation. Today, CoreWeave is the tip of the spear bringing new Nvidia architectures—H100, H200, GB200, GB300—into commercial production at scale, each generation moving from training to experiments to long-tail inference over a multi-year lifespan.


3

The «Box» Model: How CoreWeave Finances Billions in GPUs

CoreWeave securitizes client contracts, data-center leases, and GPUs into discrete cash-flow vehicles that self-liquidate.

1

Sign a 5-year contract CoreWeave negotiates a contract with a hyperscaler or foundation-model customer (e.g., Microsoft) for compute over five years.

2

Create the «box» The contract, GPU purchase orders from Nvidia, and data-center lease are placed into a special-purpose vehicle (SPV) that governs all cash flow.

3

Build and deliver compute CoreWeave constructs the infrastructure and begins delivering compute. Microsoft pays the box directly, not CoreWeave.

4

Waterfall payments Incoming revenue pays the data center, power bill, interest, and principal in sequence; excess flows back to CoreWeave equity.

5

Self-liquidation in 2.5 years Within 2.5 years of a 5-year deal, all principal and interest is repaid. The remaining 2.5 years generate pure equity returns.

6

Scale and lower cost of capital Each successful box builds lender confidence. CoreWeave has dropped its cost of capital by 600 basis points in two years, approaching hyperscaler borrowing rates.


4

GPU Depreciation Debate: «Nonsense» or Real Risk?

CoreWeave's Intrator calls obsolescence fears trader-driven FUD; A100 prices appreciated year-over-year.

The facts on the ground is they're buying it for 5 years. If people are willing to pay me for it, it still has value. The A100s, the Ampers this year, the price has appreciated through the year.

Michael Intrator, CEO, CoreWeave


5

Perplexity's Three Strategic Layers

🔍
Ask: Multi-Model Search
Users can pick any frontier model (GPT, Claude, Gemini) or let Perplexity auto-route. The answer is grounded in real-time web data, building trust through accuracy.
🌐
Comet: Agentic Browsing
An AI-controlled browser that can scrape LinkedIn, fill forms, upload files, and automate multi-step workflows—server-side or on-device.
💻
Computer: Full Orchestration
AI becomes the operating system. It manages files, spins up sandboxes, accesses connectors (Google Workspace, Slack), and delegates sub-agents across hundreds of models to complete end-to-end objectives.

6

Why Perplexity Believes Multi-Model Orchestration Wins

Models are specializing, not commoditizing—neutrality lets Perplexity take the best of each.

Aravind Srinivas argues that Perplexity's strategic advantage is being «Switzerland»—no allegiance to a single model vendor. As Anthropic CEO Dario Amodei observed, models have begun to specialize: even within coding, Claude Code and Codex have distinct strengths. Perplexity's iOS engineers prefer Codex; backend engineers prefer Claude Code. Across other domains—visual synthesis, audio, multimodal reasoning—different models excel. The company that can intelligently route workloads and orchestrate the «best musician» for each task delivers superior value without the liability of betting on one architecture.

Srinivas introduced «Model Council,» a feature that runs the same prompt across five models, then highlights where they agree, disagree, and where nuances lie—addressing the problem Jensen Huang described of manually querying five AIs and applying human judgment. Perplexity abstracts that complexity. For enterprise customers on the $400/month Enterprise Max tier, every dollar of revenue has positive gross margins because the company isn't just reselling tokens—it's selling orchestration, routing efficiency, and recurring subscriptions. Unlike wrapper companies that subsidize usage, Perplexity's multi-model harness and retrieval-augmented generation (RAG) architecture reduce context-window bloat and token costs, enabling profitability at the unit level even as the company scales.


7

Perplexity's Enterprise Push: Computer as a Slack Bot

Thousands of companies use Perplexity Computer; internal employees talk more to it than colleagues.

💡

Perplexity's Enterprise Push: Computer as a Slack Bot

Perplexity's fastest-growing revenue stream is enterprise. Computer exists as a Slackbot on the Enterprise plan, and Srinivas reports that «people are talking more to Computer on Slack than to other people.» The company has saved over $100 million in labor costs for Enterprise Max customers. One board memo, one partnership deck, and one press briefing prep were all one-shotted by Computer—tasks that previously required design teams, comms staff, and hours of human coordination.


8

Mistral's Thesis: Open Models Unlock Proprietary Data

Closed models can't integrate decades of manufacturing or financial IP; open models can.

CLOSED MODELS
Limited Post-Training & Integration
Enterprises cannot add parameters, cannot deeply customize weights, and cannot deploy on arbitrary clouds or on-premises hardware. Closed models are one-size-fits-all, disconnected from decades of proprietary signals in factories, trading desks, or scientific instruments.
OPEN MODELS
Deep Customization & Edge Deployment
Mistral's Forge platform deploys portable training infrastructure on customer premises. Forward-deployment PhD engineers spend weeks with subject-matter experts (e.g., ASML defect detection), ingest proprietary data, retrain models, and hand off ownership—enabling enterprises to leverage IP while maintaining full data sovereignty.

9

Mistral's Forward-Deployment Model

Mistral sends PhD scientists to customer sites, trains models on-premises, ensures zero data leakage.

1

Deploy portable platform Mistral's training tools and data-processing pipelines are installed on the customer's infrastructure, ensuring no data flows back to Mistral.

2

Send forward-deployment engineers PhD-level scientists work alongside vertical experts (e.g., image-scanning engineers at ASML) to understand workflows and defect-detection logic.

3

Ingest proprietary data Customer data—often decades of manufacturing signals, medical images, or financial transactions—is used to fine-tune open-source models.

4

Train and hand off Models are retrained on-site. Once complete, the customer can operate independently, retraining and iterating without Mistral's ongoing involvement.


10

IREN: Building Data Centers Where the Wind Blows

IREN co-locates compute with stranded renewable energy, avoiding transmission bottlenecks and achieving 100% renewables.

Daniel Roberts and his brother started IREN eight years ago in Sydney, betting that the digital world's exponential growth would strain real-world infrastructure. Their thesis: build large-scale data centers, bootstrap with Bitcoin mining cash flow, then layer in higher-value use cases as they emerge. Today, IREN is swapping out all Bitcoin miners for AI chips, and the company has 4.5 GW of capacity under development—nearly the annual power consumption of the entire San Francisco Bay Area.

The breakthrough insight: West Texas has 45–50 GW of wind and solar generation, but only a 12 GW transmission line to export power to Dallas and Houston. That creates stranded renewable energy. IREN builds data centers at the source, monetizing excess wind and solar into compute and exporting it «at the speed of light as tokens.» The utility underwrites variability and guarantees 24/7 reliable power; IREN pays for 100% renewables without needing to manage batteries or backup generation. Four years ago, when IREN proposed a 750 MW site in the middle of the Texas desert, the traditional data-center industry thought they were crazy. Today, the company has signed a $9.7 billion contract with Microsoft—representing just 5% of its total capacity.


11

Key Infrastructure Metrics

CoreWeave, Perplexity, and IREN shared hard numbers on contracts, capacity, and costs.

CoreWeave's Average Contract Length
5 years
Protects against demand air pockets and gives lenders confidence in cash-flow predictability.
CoreWeave Cost-of-Capital Reduction
600 basis points
Dropped borrowing costs over two years by demonstrating box-level profitability and self-liquidation.
Perplexity Monthly Active Users
Tens of millions
Consumer adoption growing rapidly; enterprise is the fastest-growing revenue segment.
Perplexity Enterprise Max Savings
>$100 million
Cumulative labor-cost savings reported by highest-tier enterprise customers using Computer.
IREN Total Power Capacity
4.5 GW
Equivalent to the annual power consumption of the San Francisco Bay Area.
IREN Microsoft Contract Value
$9.7 billion
Represents only 5% of IREN's total contracted capacity.
West Texas Renewable Generation
45–50 GW
Wind and solar capacity far exceeds the 12 GW transmission line to load centers, creating arbitrage opportunity.

12

The Real Constraint: Time to Compute

GPUs are available—but building data centers in remote Texas requires thousands of tradespeople.

IREN's Daniel Roberts argues that the industry narrative has shifted: the constraint is no longer GPU supply, it's «time to compute.» For IREN, which secured land and power eight years ago, the bottleneck is physical: breaking ground, pouring foundations, installing water-cooling systems, running fiber, and hiring electricians, HVAC technicians, and construction crews. These are hard manual-labor jobs in remote locations, often requiring workers to live in temporary housing for multi-month tours of duty.

IREN's policy is to hire locally first, expanding the radius in 20-mile increments as needed. They target areas where heavy electrical infrastructure already exists—typically where old manufacturing has shut down—and retrain local workforces. The company is partnering with trade schools and universities to build a new generation of skilled labor. Starting salaries for electricians and construction workers on IREN sites are in the $150K range, making trade careers financially competitive with white-collar knowledge work. Roberts sees this as a reversal of the decades-long offshoring of manufacturing: AI infrastructure is bringing high-paying blue-collar jobs back to rural America, powered by the same wind and solar farms that replaced coal and gas in the energy transition.


13

Jevons's Paradox and the Infinite Demand for Compute

As inference gets faster and cheaper, humans will generate exponentially more queries—not fewer.

💡

Jevons's Paradox and the Infinite Demand for Compute

Roberts invoked Jevons's paradox: if generating an image in ChatGPT today takes two minutes, users carefully craft prompts and generate sparingly. If a 10× increase in compute drops that to 5–10 seconds, will users generate more or fewer images? Many more. The same applies to autonomous vehicles, real-time translation, and agentic workflows. Faster, cheaper inference doesn't reduce demand—it unlocks entirely new use cases and consumption patterns, creating a positive feedback loop that drives exponentially higher compute requirements.


14

Títulos mencionados

NVDANvidia Corporation
MSFTMicrosoft Corporation
GOOGLAlphabet Inc.
METAMeta Platforms Inc.
IRENIREN (Iris Energy Limited)

15

Pessoas

Jason Calacanis
Host & Interviewer
host
Michael Intrator
CEO & Co-Founder, CoreWeave
guest
Aravind Srinivas
CEO & Co-Founder, Perplexity
guest
Arthur Mensch
CEO, Mistral AI
guest
Daniel Roberts
Co-CEO & Co-Founder, IREN
guest
Jensen Huang
CEO, Nvidia
mentioned
Mustafa Suleyman
Founder, Inflection AI
mentioned
Dario Amodei
CEO, Anthropic
mentioned
Sam Altman
CEO, OpenAI
mentioned
Elon Musk
CEO, SpaceX / Tesla / xAI
mentioned

Glossário
Box (SPV)A special-purpose vehicle that securitizes a client contract, GPU purchase, and data-center lease into a discrete cash-flow waterfall, enabling non-recourse financing.
Time to ComputeThe bottleneck in AI infrastructure: not GPU availability, but the time required to build data centers, hire tradespeople, and energize facilities.
Jevons's ParadoxAn economic principle: as efficiency increases and cost per unit falls, total consumption rises rather than falls, because new use cases become viable.
Forward-Deployment EngineeringMistral's model of sending PhD-level engineers to customer sites to train models on proprietary data, ensuring zero data leakage and deep domain customization.
Model CouncilPerplexity feature that runs the same prompt across multiple models, then highlights agreements, disagreements, and nuances—automating multi-model comparison.

Aviso: Este é um resumo gerado por IA de um vídeo do YouTube para fins educacionais e de referência. Não constitui aconselhamento de investimento, financeiro ou jurídico. Verifique sempre as informações com as fontes originais antes de tomar decisões. O TubeReads não é afiliado ao criador do conteúdo.