The VC Corner

Jensen Huang: 10 Lessons From the CEO Building the Most Important Company in History

RD

Ruben Dominguez (The VC Corner)

Apr 10, 2026

16 min read

Jensen Huang: 10 Lessons From the CEO Building the Most Important Company in History

Source: The VC Corner · Author: Ruben Dominguez (The VC Corner) · Date: 2026-04-10 · Original article

This is a teaching summary of Ruben Dominguez's distillation of a two-hour Lex Fridman interview with Jensen Huang, the founder and CEO of NVIDIA. Dominguez watched the full interview and pulled out the ten ideas that, in his view, every founder, investor, and operator should internalize about how Jensen builds, communicates, and thinks about AI infrastructure. Below, each lesson is explained in plain English, with the analogies, examples, and reasoning preserved so you walk away with the same intuition the original article teaches.


1. NVIDIA Is Not a Chip Company — It Is a Computing Platform ($4 Trillion of Distinction)

The single most important reframe in the piece: Jensen never set out to build the best GPU. He set out to build a computing platform. A GPU is a product; a platform is an environment that other companies build their own products on top of (think iOS or Windows, but for math-heavy computation).

Why this matters strategically: an "accelerator" is a specialty chip that speeds up one narrow kind of work — say, graphics for video games. The market for any single accelerator is, by definition, capped at the size of that one market. A computing platform, on the other hand, can absorb every workload that involves computing — gaming, scientific simulation, training AI, running AI, robotics, drug discovery, weather forecasting. The total addressable market is essentially "everything that computes."

Jensen described the path NVIDIA walked, deliberately, step by step:

  • Programmable pixel shader — the first move that let developers write small programs that ran on the GPU, instead of using fixed graphics features. This was the first crack of general programmability.
  • FP32 in shaders — adopting the IEEE 32-bit floating point standard inside graphics shaders. In plain English: the GPU started doing math the way scientific software expected it to be done. That single decision quietly invited non-gaming developers (physicists, researchers) to start using GPUs.
  • CUDA on GeForce — putting CUDA (NVIDIA's general-purpose programming layer for GPUs) on every consumer gaming card. This was the bet that almost killed the company and defined everything that came after.

Jensen's framing: every platform company in history has had to find the narrow path between too specialized (your market caps out) and too general (you lose your edge against specialists). Finding and walking that ridge is the whole game.

For founders and investors, this is why "platform" companies command much higher valuation multiples than pure hardware plays — investors are pricing in TAM expansion, not just current revenue. If your pitch positions you as the next great accelerator for one workload, you've capped your own ceiling.


2. The CUDA Bet Almost Destroyed NVIDIA — That Is Why It Defined NVIDIA

Few decisions in corporate history have been vindicated as completely as putting CUDA on every consumer GeForce card. At the time, it nearly bankrupted the company.

In Jensen's words: "CUDA increased our cost of that GPU so tremendously it completely consumed all of the company's gross profit dollars. Our market cap went down to like $1.5 billion. We clawed our way back slowly, but we carried CUDA on GeForce."

The reason this looks insane on a spreadsheet but brilliant in hindsight is one of the deepest lessons in tech strategy: install base defines an architecture — not elegance, not benchmarks.

Jensen's proof point is the x86 instruction set — the architecture inside almost every PC and server CPU. By any engineering criterion, x86 is one of the least elegant designs in computing history. Beautiful, clean RISC architectures (designed by world-class engineers) tried to displace it for decades and largely vanished. x86 won because it was already everywhere. Software was written for it; therefore, more software was written for it; therefore, more hardware was built for it.

The CUDA bet had three parts:

  1. Subsidize — eat the cost and put CUDA on every GeForce card so any researcher building their own PC at home would, by default, have a CUDA-capable machine.
  2. Evangelize — go to universities, write textbooks, teach classes, fund students. Get an entire generation of developers building on CUDA.
  3. Wait — trust that with a vast, idle, programmable parallel-compute install base sitting in the world, something amazing would eventually emerge that needed exactly this.

That something was deep learning. By the time AlexNet showed in 2012 that GPUs could train neural networks dramatically faster than CPUs, CUDA was already the default — every researcher who wanted to try the new ideas already had the tools, the libraries, and the hardware.

Lesson: distribution moats are not built in a quarter. They are built across a decade of painful, unprofitable commitment. If your strategy depends on an "if we build it, they will come" install-base bet, the question is not whether the bet is risky — it is whether you can survive long enough for the payoff.


3. Inference Is Thinking — and Thinking Is Hard

One of the costliest mispredictions in the AI industry was the belief that inference (the act of an already-trained AI model answering a query) would be cheap, simple, and quickly commoditized — handled by tiny, low-power chips. Jensen publicly said the opposite for years. Most of the industry didn't listen.

His framing, in his own words: "Inference is thinking. Thinking is way harder than reading. Pre-training is just memorization. Thinking, reasoning, solving problems, taking new experiences and decomposing them into solvable pieces. How could that possibly be compute light?"

Translation for non-AI people:

  • Pre-training is when a model reads a huge pile of text and absorbs patterns. It's roughly "memorize the world."
  • Inference, in modern reasoning models, is the model actively working through a new problem at the moment you ask it — breaking the problem down, trying approaches, checking itself. It's closer to thinking than to looking up an answer in a book.

If thinking is harder than reading, why on earth would the chip that does the thinking be smaller and cheaper than the chip that did the reading?

Jensen lays out four scaling laws that, taken together, explain why compute demand keeps growing rather than tapering off:

  1. Pre-training scaling — bigger models trained on more data continue to get smarter. Classic scaling, still alive.
  2. Post-training scaling — after the initial training, you keep improving the model with techniques like synthetic data generation (the model helps create more training material for itself) and reinforcement learning from feedback. This phase has its own compute curve.
  3. Test-time (inference) scaling — at the moment of answering, modern models reason through a problem, often producing long internal chains of thought before responding. The longer they "think," the better the answer. By definition, this is compute-intensive.
  4. Agentic scaling — an agent can spawn sub-agents, which spawn their own sub-agents. One question becomes a team of AIs working in parallel; the team becomes an army. Each of those agents is itself doing test-time reasoning.

Each of these laws is independently consuming compute. They stack. That is why "inference will be cheap" was wrong — and why NVIDIA bet the other way and is being vindicated.

The takeaway for builders: if you are designing AI agents, the compute question isn't a footnote. It's the foundation of your unit economics.


4. The Computer Just Changed from a Warehouse to a Factory

This is Jensen's clearest mental model for what is actually happening, and why it justifies his belief that NVIDIA's market is essentially unbounded.

In his words: "Computers, because they were a storage system, were largely a warehouse. We're now building factories. Warehouses don't make much money. Factories directly correlate with a company's revenues."

Unpack the analogy:

  • The old computer was a warehouse. You stored data in it (web pages, records, files), and when someone asked, you went and retrieved the right pre-existing item. The economics of warehouses are well-known — they're a cost center.
  • The new computer is a factory. You feed in raw materials (a question, some context) and it manufactures a brand-new output — an answer, a piece of code, an image, a plan — calibrated specifically to that situation. That output didn't exist a moment ago. Factories produce goods that have prices; they're a profit center.

The "goods" the AI factory produces are tokens (the small chunks of text or other data a model emits). And those tokens already have segmented pricing in the real market:

  • Free tokens for casual chat use.
  • Premium tokens for specialized professional tasks.
  • High-value tokens at roughly $1,000 per million for the most demanding reasoning workloads.

This reframing rewires entire business models. SaaS financial models built in 2022 don't have a row for token-based revenue. MRR projection sheets need to add token throughput. Every enterprise that runs a "token factory" is running a revenue-generating machine — and the world will need many more of them, which is the market NVIDIA is selling shovels into.


5. Shape Belief Systems Continuously — So Announcements Feel Inevitable

Most leaders make a decision privately, then announce it. The announcement is a surprise to most of the audience, who then have to be convinced.

Jensen does the inverse. "I like to imagine that when I announce these things, the employees are saying, 'Jensen, what took you so long?' I've been shaping their belief system for some time. On the day I declare it, there's a hundred percent buy-in."

He runs this play at every layer simultaneously:

  • Board — continuously briefed on business conditions, growth drivers, and the shape of the market. No big update is ever truly new.
  • Management team — instead of deciding alone and informing, he reasons through problems in front of them. By the time a conclusion forms, everyone watched it form.
  • Partners and supply chain — his GTC keynotes are not just product announcements; they are coordination signals to the ~200 supplier CEOs whose roadmaps need to align with NVIDIA's. The keynote tells them what to build for the next two years.
  • Industry and public — public statements about the future essentially manifest it. Once a credible CEO says "this is happening," competitors and customers begin acting as if it's true, which makes it more true.

The goal is never surprise. Surprise means you failed to bring people along. The goal is inevitability — the feeling that the announced thing was already the obvious next step.

For founders raising capital: the investors who write the biggest checks are the ones who already felt the inevitability before the pitch deck arrived. Shape the belief first. The ask comes second.


6. The Moat Is Not a Chip — It Is a Platform 43,000 People and Millions of Developers Built Together

When asked directly about NVIDIA's moat, Jensen does not mention hardware specs. He mentions trust.

"Our single most important property as a company is the install base of our computing platform. It wasn't three people that made CUDA successful. It was 43,000 people. And the several million developers who trusted that we were going to continue to make CUDA 1, 2, 3, ... 13. You could take that to the bank."

Look at the calculation from a developer's seat:

  • If I commit my project to CUDA today, I know that in six months CUDA will be roughly 10× better because NVIDIA's velocity is relentless.
  • I also know my CUDA code will run on every cloud, in every country, across every industry — instantly, simultaneously.

That combination — relentless improvement velocity, universal reach, and 30 years of kept promises — is what no competitor can replicate by buying or copying. It can only be built, slowly, over decades.

This is the honest answer to every due-diligence question about defensibility. Increasingly, sophisticated AI investors score distribution and developer trust higher than raw model performance when valuing early-stage AI companies. If your pitch deck leads with benchmark scores and buries the ecosystem slide, flip the order.


7. The Org Chart Should Mirror the Product — 60 Direct Reports and No One-on-Ones

Most company org charts look the same regardless of what the company actually builds. Jensen thinks that's insane, because the structure of the organization shapes what the organization is capable of building.

"My direct staff is 60 people. I don't have one-on-ones because it's impossible. No conversation is ever one person. We present a problem and all of us attack it, because we're doing extreme co-design and literally the company is doing extreme co-design all the time."

A few terms worth unpacking:

  • Extreme co-design means designing every component (chip, memory, networking, software, packaging, cooling) together and simultaneously — because every component changes the constraints on every other component. You can't optimize the GPU in isolation if its performance depends on the memory bandwidth and the optical interconnects between racks.
  • 60 direct reports with no one-on-ones is the organizational shape that this product philosophy requires. Every problem is presented to the whole room — memory experts, GPU architects, networking specialists, optical engineers — and attacked together. There is no point in pulling each leader aside individually because no real problem belongs to just one of them.

The principle for early-stage founders: hire for the product you are actually building, not the org chart template you inherited from your last job. Your headcount plan, your operating plan, and your product architecture should all reflect the same structure. If they're misaligned, investors doing diligence will find the seam before you do.


8. Use the "Speed of Light" as the Benchmark for Every Engineering Decision

Continuous improvement — make today's process 3% faster than yesterday's — is not how Jensen approaches engineering problems. He strips a problem down to its physical limits first.

"Everything that we do is compared against the speed of light. Memory speed, math speed, power, cost, time, effort, number of people, manufacturing cycle time. I force everybody to think about the physical limits for everything before we do anything."

The example he gives is concrete and worth sitting with. Imagine a process today takes 74 days. Someone walks in proudly proposing an optimization to bring it down to 72 days. Jensen is not interested. He wants to know: what do first principles say this process should take if every step ran at its physical or theoretical maximum? The answer is often 6 days.

Now the entire conversation changes. Instead of negotiating from 74 down to 72, you are negotiating from 74 down to 6. That gigantic gap is where every assumption, every legacy constraint, and every "we do it this way because we always have" hides. You can only see those hidden assumptions if you anchor against the speed-of-light limit, not against last quarter's number.

Continuous improvement optimizes the existing path. Speed-of-light thinking asks whether the path should exist at all.

The article extends the analogy to AI workflows: the old prompt-engineering mindset was iterative — tweak, rerun, improve marginally. Newer "context engineering" disciplines first ask what the theoretical ceiling of an AI interaction is, then work backward from there. Same logic, applied to a different domain.


9. Intelligence Is a Commodity. Humanity Is Not. (Jensen's Most Important Idea)

This is the single idea Dominguez calls Jensen's most important — and it cuts directly against the most common public anxiety about AI.

Jensen's example: "The purpose of a radiologist is to diagnose disease and help patients. Because we're able to study scans so much faster now, you could study more scans. You could diagnose better. We now have a shortage of radiologists in the world. The amazing thing is it's so obvious this was going to happen."

Recall the timeline: by 2019, computer vision systems already exceeded average human performance at reading certain medical scans. The widely repeated prediction was that radiologists were about to be automated out of existence. Instead, the number of radiologists grew, and there is now a shortage of them.

The alarmist prediction was wrong because it confused the task with the purpose:

  • The task of a radiologist is reading scans. AI now helps read scans much faster.
  • The purpose of a radiologist is to diagnose disease and care for patients. When the task gets faster, the purpose expands: more scans get read, finer findings get caught, more patients get helped. Demand for the purpose goes up, not down.

The same dynamic is now playing out in software engineering:

  • Task: writing lines of code. AI does much of this now (recent industry data shows roughly 75% of programming tasks are AI-assisted at leading AI companies).
  • Purpose: deciding which problems are worth solving, evaluating whether a result is correct, connecting dots across domains, finding new problems to attack, innovating. AI does not replace this — it amplifies it. Headcount of engineers at top AI companies is growing.

Concrete real-world illustration from the article: an agency owner using AI agent pipelines now charges clients $50,000 for engagements they used to bill $500/month for. They are not doing more tasks. They are operating at a higher level of purpose — designing systems, owning outcomes — while AI handles the underlying tasks.

Strategic takeaway for AI product builders: tools that merely replace tasks lose to commoditization (any competitor can offer the same task replacement). Tools that expand purpose build retention — and retention is what investors put a premium valuation on. Run your CLTV/CAC numbers across both categories and the gap is not marginal.


10. China Is the Fastest-Innovating Country in the World Right Now

Most Western executives hedge when asked about China. Jensen does not. His analysis is structural, not political.

"50% of the world's AI researchers are Chinese, plus or minus. They have insane competition internally. And what remains is an incredible company."

He breaks down the structural reasons:

  • Talent density — strong math and science education at scale produces a large pool of world-class engineers. Half of all AI researchers globally are Chinese.
  • Internal competition — multiple provinces and city governments actively compete to attract and incubate companies. This produces more startups, more iteration cycles, and after the brutal selection pressure, better survivors. It is essentially a Darwinian funnel run at the scale of a country.
  • Cultural openness around knowledge — the article observes that, as a social norm, sharing technical knowledge among peers is the default; open source is treated as normal, not as a strategy. This creates faster diffusion of techniques.
  • Builder culture at the top — most senior Chinese tech leaders are engineers by background. Most senior American business leaders are lawyers or financiers. Engineer-CEOs make different bets than lawyer-CEOs.

For investors, this is not a geopolitical observation — it's a deal-flow observation. Cross-border AI deals are accelerating, and recent YC batches already include multiple companies with Chinese founding teams raising from US angels.


The NVIDIA Playbook — What to Steal

Jensen has been building the same company for 33 years. The chips changed. The market grew by roughly seven orders of magnitude. The underlying philosophy never changed.

A summary of who should take what:

  • For founders: the narrow path between specialization and generalization is where every great platform lives. Find it and walk it one step at a time, even when each step looks unprofitable.
  • For investors: the single most important diagnostic question for a tech company is whether it is competing for share in an existing market or creating a new one. Share games have ceilings; creation games do not. The difference is generational, not marginal.
  • For operators: every process has a speed of light. Find that limit before you start optimizing. Otherwise you will spend years getting from 74 days to 72 when the real answer was 6.

Five principles to internalize:

  1. Install base defines a platform — not elegance, not benchmarks, not marketing.
  2. Shape belief systems continuously — by the time you announce, it should already feel inevitable.
  3. The org chart should mirror the product — extreme co-design requires extreme co-management.
  4. Intelligence is a commodity. Humanity is not. — purpose and task are not the same thing.
  5. Test everything against the speed of light — continuous improvement quietly optimizes the wrong thing.

The computer just became a factory. Factories make money. Jensen figured this out 30 years ago.

#AI#AI_AGENTS#ENGINEERING#PRODUCT#GROWTH#STARTUPS

Author

Ruben Dominguez (The VC Corner)

The weekly builder brief

Subscribe for free. Get the signal. Skip the noise.

Get one focused email each week with 5-minute reads on product, engineering, growth, and execution - built to help you make smarter roadmap and revenue decisions.

Free forever. Takes 5 seconds. Unsubscribe anytime.

Join 1,872+ product leaders, engineers & founders already getting better every Tuesday.