Jensen Huang: Why Nvidia’s Monopoly on AI Compute Is Hard to Break


Aorus computer hardware with intricate design.

At the 2024 CES keynote, Jensen Huang walked out in his signature black leather jacket to a crowd that treated him like a rock star. The chips he was announcing — Blackwell architecture, next-generation data center GPUs — would go on to generate more revenue in a single quarter than most Fortune 500 companies make in a year. Nvidia posted $35.1 billion in Q3 FY2025 revenue. The data center segment alone: $30.8 billion. In one quarter. That’s not a typo, and it’s not because Jensen got lucky. It’s because he made a series of bets over more than a decade that almost everyone else thought were wrong — and they turned out to be the exact infrastructure the AI revolution needed.

If you want to understand where AI is going, you need to understand Jensen Huang. Not just as a CEO, but as a strategic thinker who has shaped what AI physically runs on, how companies think about compute, and increasingly, what the next wave of AI — physical AI, robotics, agentic systems — is going to look like. He’s not a researcher. He’s not building frontier models. But without him, the frontier wouldn’t exist.

The CUDA Bet: The Decision That Changed Everything

In 2006, Nvidia released CUDA — a parallel computing platform that let developers use Nvidia GPUs for general-purpose computation, not just rendering graphics. At the time, this looked like a niche developer tool. Jensen invested heavily in it anyway, for years, with no obvious killer app in sight. The gaming business was subsidizing what looked like an expensive science project.

Then, in 2012, a paper called “ImageNet Classification with Deep Convolutional Neural Networks” showed up. You probably know it as the AlexNet paper. Geoff Hinton, Ilya Sutskever, and Alex Krizhevsky used two Nvidia GTX 580 GPUs to train a neural network that demolished the competition in image recognition. They used CUDA to do it. The AI research community noticed immediately.

What followed was a decade-long compounding effect. Every major AI lab — Google Brain, OpenAI, DeepMind, Meta AI — built their research infrastructure on Nvidia GPUs running CUDA. When the transformer architecture took off after 2017, training those models required massive parallel compute. Nvidia had it. Competitors didn’t — at least not with the software ecosystem that made it usable. AMD has better hardware specs on paper in some cases, but CUDA’s decade-plus head start in developer tooling, libraries, and institutional knowledge is an enormous moat. Andrej Karpathy, who trained neural nets at Stanford and later led Tesla’s Autopilot and worked at OpenAI, has talked repeatedly about how CUDA proficiency is a core skill for serious AI researchers — it’s that embedded in the stack.

Jensen didn’t predict LLMs specifically. But he bet that parallel compute would matter for everything, and he built the tools that made that compute accessible. That’s the move.

How Jensen Thinks About AI: Physical AI and the Next Platform Shift

If you listen to Jensen’s keynotes and interviews — and he gives a lot of them, often at length — a few consistent themes emerge. He doesn’t talk much about AGI timelines or consciousness. He talks about platforms, infrastructure, and what he calls “physical AI.”

His thesis, stated repeatedly at events like GTC 2024 and in interviews with outlets like Acquired, is roughly this: the first wave of AI was about perception and understanding — classifying images, transcribing audio, generating text. The next wave is about AI that operates in the physical world. Robots. Autonomous vehicles. Industrial automation. And that requires a fundamentally different kind of AI — one that understands physics, can plan in 3D space, and can be trained in simulation before being deployed on hardware.

This is why Nvidia built Omniverse — a platform for simulating physical environments at scale. And it’s why Isaac, Nvidia’s robotics platform, matters strategically. Jensen has framed the humanoid robot moment as the next computing platform shift, analogous to the PC or smartphone. Companies like Figure AI, Boston Dynamics, and Agility Robotics are building on Nvidia hardware and software stacks. Whether humanoid robots become a mass-market product in the next five years is genuinely uncertain — Jensen himself has said timelines are hard to predict — but Nvidia is positioning to be the infrastructure layer regardless of which hardware form factor wins.

Nvidia also released Cosmos in early 2025 — a world foundation model designed to generate physically realistic simulated environments for training robots and autonomous systems. It’s not a consumer product. It’s infrastructure for labs and companies that need synthetic training data at scale. Jensen’s consistent play: own the picks and shovels, not the gold.

The Blackwell Architecture and What It Actually Means for AI Labs

When Nvidia announced the Blackwell GPU architecture in 2024, the specs were genuinely significant — not because of the raw numbers, but because of what they enable at scale. Blackwell chips are designed to handle the inference workload of large language models more efficiently, which matters because inference (running the model, not training it) is where most of the actual cost sits once a model is deployed.

For context: OpenAI, Anthropic, Google, and Meta are spending billions on GPU clusters. Microsoft’s investment in OpenAI is partly about securing Azure data center capacity. The chip supply chain for AI is now a geopolitical issue — the US government’s export controls on advanced chips to China are directly about Nvidia’s H100 and A100 chips. When Jensen says Nvidia is supply-constrained and demand is extraordinary, he’s not doing investor relations theater. Every major cloud provider is in an arms race to deploy more Nvidia silicon faster than their competitors.

The GB200 NVLink rack — Blackwell’s data center configuration — connects GPUs with extremely high-bandwidth interconnects so they can operate as a unified compute system rather than individual chips. This matters for training frontier models, which require coordinating computation across thousands of GPUs simultaneously. Jensen describes this as building “AI factories” — not servers, but purpose-built infrastructure for AI production at scale. The language shift is intentional and reflects how he thinks about Nvidia’s market: not selling components, but selling the capacity to produce intelligence.

Jensen vs. The Other Tech Titans: A Quick Framework

It’s useful to understand how Jensen’s position differs from the other major figures shaping AI right now, because they’re often conflated but are doing very different things.

Ty Sutherland

Ty Sutherland is the Chief Editor of AI Rising Trends. Living in what he believes to be the most transformative era in history, Ty is deeply captivated by the boundless potential of emerging technologies like the metaverse and artificial intelligence. He envisions a future where these innovations seamlessly enhance every facet of human existence. With a fervent desire to champion the adoption of AI for humanity's collective betterment, Ty emphasizes the urgency of integrating AI into our professional and personal spheres, cautioning against the risk of obsolescence for those who lag behind. "Airising Trends" stands as a testament to his mission, dedicated to spotlighting the latest in AI advancements and offering guidance on harnessing these tools to elevate one's life.

Recent Posts

Person Primary Role in AI Key Bet Biggest Risk
Jensen Huang (Nvidia) Infrastructure / Hardware Own the compute layer for every AI approach Custom silicon from Google (TPUs), AWS (Trainium), Meta catching up
Sam Altman (OpenAI) Frontier model development Scale to AGI; dominate the application layer Competition from Anthropic, Google, open-source models
Demis Hassabis (Google DeepMind) Research + frontier models Scientific AI — AlphaFold, Gemini, long-horizon reasoning Organizational complexity; Google’s monetization pressure