NVIDIA is the company that made the AI revolution physically possible. Every time you ask ChatGPT a question, generate an image with DALL-E, or build an app with Taskade Genesis, the computation runs on NVIDIA hardware. Over 90% of all AI training in the world happens on NVIDIA GPUs.
But NVIDIA didn't start as an AI company. It started as three engineers in a Denny's booth with a bet on 3D graphics — and a CEO who has led it for over 32 years through near-bankruptcy, a market cap crash, and one of the most dramatic pivots in business history. This is the complete story. 🧬
TL;DR: NVIDIA went from a 1993 Denny's founding to a $3.4T+ company powering 90%+ of AI training. Jensen Huang's CUDA bet seeded the entire AI revolution. As of May 2026, he frames AI as a shift from retrieval to generative computing with physical AI (robo-taxis, humanoid robots) right around the corner. Try frontier AI models in Taskade →
Jensen Huang — Founder and CEO of NVIDIA. From a Denny's booth in 1993 to the most valuable company in the world.
🤖 What Is NVIDIA?
NVIDIA Corporation was founded on January 25, 1993, by Jensen Huang, Chris Malachowsky, and Curtis Priem at a Denny's restaurant in San Jose, California. The name derives from "invidia," the Latin word for envy. As of March 2026, NVIDIA is the most valuable public company in the world at over $3.4 trillion, powering more than 90% of all AI model training with its GPU architecture and CUDA platform.
"I like to describe our company's journey with a vision. And the vision was this simple belief: Accelerated computing was going to be an important discipline."
Jensen Huang, GTC 2024 Keynote
What started as a graphics chip company for PC gaming has evolved into the most valuable company in the world. As of March 2026:
- Market capitalization: Over $3.4 trillion — surpassing Apple, Microsoft, and every other public company
- Revenue: $130.5 billion in fiscal year 2025 (ending January 2026)
- Data center revenue: Over $115 billion annually — dwarfing the gaming business that built the company
- CUDA developers: 5+ million worldwide using NVIDIA's computing platform
- AI market share: Over 90% of all AI model training runs on NVIDIA hardware
- Employees: Over 32,000 globally across 30+ countries
The company's product lineup spans GPUs for gaming (GeForce), AI training and inference (H100, H200, Blackwell, Vera Rubin), autonomous vehicles (DRIVE), robotics (Isaac, Cosmos), and enterprise AI software (NVIDIA AI Enterprise, NeMo, NemoClaw). But the through-line from 1993 to 2026 is the same bet Jensen Huang made at Denny's: parallel computing will eat the world.
🧬 From Retrieval to Generative Computing — A New Paradigm
The single most important framing Jensen Huang offered in his May 2026 SCSP interview is that AI represents a fundamental shift in how computing works. For 60 years, computing was retrieval-based: people created content (images, videos, articles, ads, product pages), stored it in data centers (centers of data), and software fetched it back based on cookies, queries, or recommender systems. Search returned ranked links. Shopping returned one of sixteen pre-made ads. YouTube fed you whatever it had. Every output existed before you arrived.
Modern AI is generative computing. Based on your context, intent, and the live state of the world, the system perceives information in many forms (text, images, video), reasons about what you actually need, plans an answer, and generates a response that has never existed before. The output is not a stored file but a stream of numbers — tokens — reformulated into text, images, audio, code, video, or robotic action.
"It used to be retrieval-based. Now it's generative. Every time you use computing, based on your context, based on the changes of ground truth, based on your intention, intelligence will be applied and provide you the information that best suits you."
Jensen Huang, SCSP "Memos to the President" (May 2026)
The economic implication is enormous: instead of a center of data, AI requires a center of computation. Every interaction generates novel tokens, and tokens are produced by GPUs running for milliseconds to minutes per request. NVIDIA is positioned exactly where the industry needs it because the world is rebuilding its $1+ trillion data center fleet around generative computing — not retrieval — and that fleet runs on accelerated compute.
This is also why GTC 2026 framed AI factories as token factories: "This is how intelligence is made. A new kind of factory, generator of tokens, the building blocks of AI." Tokens harvest clean energy, unlock the secrets of the stars, train robots in virtual worlds, and clear paths in the physical one — all from the same generative pipeline NVIDIA builds.
NVIDIA CEO Jensen Huang — full GTC 2026 keynote (March 2026). The "tokens are the building blocks of AI" framing, Vera Rubin reveal, and the formal NemoClaw + OpenShell launch.
🍰 The Five-Layer AI Cake
In the May 2026 SCSP interview, Jensen Huang gave the clearest framework yet for how the AI industry is structured — and where the United States stands at each layer. He calls it the five-layer cake:
| Layer | What It Includes | US Position (May 2026) |
|---|---|---|
| 5. Adoption | Enterprise + government deployment, productivity gains, public sentiment | Most at risk — "AI fear" is hurting US adoption while Asia embraces |
| 4. Models | LLMs, biology models, chemistry, physics, animatronics, robotics | Ahead but China is close behind with extraordinary AI researcher density |
| 3. Infrastructure | AI factories, ISP/cloud platforms, deployment software | Leading — hyperscaler buildout dwarfs every other country |
| 2. Chips | GPU/CPU silicon, packaging, full systems | Dominant — NVIDIA holds 90%+ AI training share |
| 1. Energy | Power generation, grid modernization, fuel mix | Behind — China leads in production and energy tech |
Jensen's argument is that America must compete in every layer — and the layer he is most worried about is the top one. AI applications and adoption are where productivity, prosperity, and economic leadership materialize. "We were the front runners of applying technology in the last industrial revolution. We need to be careful not to be the last in this industrial revolution." Tokens at every layer — and they're not just text. AI represents biology, chemicals, physics, articulation, animatronics. Industries adjacent to chatbots (drug discovery, energy generation, heavy equipment) often don't get the voice that AI labs do, but they are critical to the future of every nation.
🏭 The Reindustrialization Thesis: Three New Plant Types
The other thesis Jensen has begun pushing publicly in 2026 is that AI is the first market force in a generation powerful enough to reindustrialize the United States. The argument: for decades America became a country where you could not get ahead without a four-year degree, a master's, or a PhD. That is unfortunate and unnecessary. AI changes the equation by creating demand for three categories of physical plants — and the high-skill, high-pay manufacturing jobs that come with them.
NVIDIA put real money behind the thesis: a half-trillion-dollar consumption commitment to bring the supply chain from East to West so chip plants, packaging plants, and computer plants can be built and operated in the United States. Jensen frames the total as "trillions of dollars of manufacturing, high-skilled labor jobs."
The bottleneck is not capital. It is energy. You cannot transform atoms (build chips, assemble racks, run AI factories) without breaking and forming covalent bonds, which requires enormous power. Jensen's pragmatic proposal: modernize the grid rather than build it from scratch. The US grid is over-provisioned to handle the worst 12 days of the year, which means the rest of the time enormous excess capacity goes unused. Service-level agreements with utilities could allow AI factories to consume that excess power on flexible terms, with on-site backup generation (solar, nuclear, or other sustainable sources) for the ~12 high-demand days.
"We have an opportunity right now with this incredible market-driven force to use this opportunity to one — make sure that the United States becomes a manufacturing nation again, create enormous amounts of high-skilled and high-paying jobs. And the second is to use this opportunity to enhance the energy system of our country."
Jensen Huang, SCSP "Memos to the President" (May 2026)
🥚 The History of NVIDIA
NVIDIA's 30-year journey from a three-person startup to a $3.4 trillion company spans five distinct eras: the GPU invention (1993–2006), the CUDA computing platform (2006–2012), the deep learning discovery (2012–2016), the AI training monopoly (2016–2022), and the generative AI explosion (2022–2026). Each era built on the last, compounding NVIDIA's technical moat until it became irreplaceable infrastructure.
Before GPUs: The State of Computer Graphics (1960s–1993)
Long before NVIDIA existed, computer graphics was a niche discipline. In 1963, Ivan Sutherland created Sketchpad at MIT — the first interactive computer graphics program. By the 1970s, companies like Evans & Sutherland built expensive graphics workstations for the military and Hollywood.
The real inflection came in the 1980s with Silicon Graphics Inc. (SGI), which made 3D graphics accessible to film studios (Jurassic Park's dinosaurs ran on SGI machines) and research labs. But SGI workstations cost tens of thousands of dollars. The PC, meanwhile, could barely render a smooth 2D sprite.
The gap between professional graphics and consumer hardware was enormous. And that gap was the opportunity three engineers saw in 1993.
The Denny's Founding (1993–1997)
Jensen Huang was 30 years old and working as a director at LSI Logic, a semiconductor company. Chris Malachowsky and Curtis Priem were engineers at Sun Microsystems. The three met regularly at a Denny's in San Jose to discuss a shared conviction: the PC would eventually need dedicated hardware for 3D graphics, and whoever built it would own the future of computing.
On January 25, 1993, they incorporated NVIDIA with $40,000 in the bank.
Jensen Huang's background shaped NVIDIA's DNA from day one. Born in Tainan, Taiwan, in 1963, he moved to the United States as a child. At age nine, his parents sent him and his brother to live with relatives in Washington state, where a mix-up landed them in a rural reform school. Huang has described mopping floors and living among troubled teenagers as a formative experience — one that taught him resilience before he was old enough to drive.
He earned a bachelor's degree in electrical engineering from Oregon State University and a master's from Stanford. Before NVIDIA, he worked at AMD (designing microprocessors) and LSI Logic (running a business unit). Both experiences gave him fluency in chip design and the business side of semiconductors.
NVIDIA's first product, the NV1 (1995), was a commercial failure. It used a non-standard rendering approach (quadratic texture mapping instead of the triangle-based rendering that the industry was standardizing around). The NV1 shipped in the Diamond Edge 3D graphics card and supported Sega Saturn game ports, but developers overwhelmingly preferred the competing 3Dfx Voodoo, which used conventional triangle rendering.
The company nearly died. NVIDIA had to lay off most of its employees and was down to its last few months of funding.
"We built a different chip at the different time for a different market. Our approach was just different enough to be incompatible, but not different enough to be better."
Jensen Huang on the NV1 failure
The turnaround came with the RIVA 128 in 1997 — NVIDIA's first product built on industry-standard triangle rendering. It sold over 1 million units in its first four months, saved the company from bankruptcy, and established NVIDIA as a serious player in the PC graphics market.
The GPU Revolution (1999–2005)
On August 31, 1999, NVIDIA launched the GeForce 256 and marketed it as the world's first GPU — Graphics Processing Unit. The term was NVIDIA's invention, and it stuck. The GeForce 256 could perform 10 million polygons per second and included hardware transform and lighting (T&L), offloading work that previously ran on the CPU.
The same year, NVIDIA went public on the Nasdaq (ticker: NVDA) at $12 per share, raising $42 million in its IPO.
What followed was a decade of dominance in PC gaming:
| Year | Product | Significance |
|---|---|---|
| 1999 | GeForce 256 | First GPU, hardware T&L |
| 2000 | GeForce 2 | Programmable pixel shaders |
| 2001 | Xbox GPU | Microsoft chose NVIDIA for the original Xbox |
| 2002 | GeForce FX | DirectX 9, 130 nm process |
| 2004 | GeForce 6 | First unified shader architecture hints |
| 2005 | GeForce 7 | SLI (dual-GPU) support |
The Xbox deal in 2001 was a milestone — Microsoft chose NVIDIA to build the graphics chip for its first gaming console, validating NVIDIA's technology at the highest level. But the relationship soured over pricing disputes, and Microsoft switched to ATI (later acquired by AMD) for the Xbox 360.
Competition with ATI Technologies (acquired by AMD in 2006 for $5.4 billion) defined this era. The two companies traded blows with each generation, driving rapid innovation. NVIDIA's key advantage was its pace of iteration — Jensen Huang pushed for a new architecture roughly every two years, a cadence he called "Huang's Law" — the observation that GPU performance for key workloads doubles approximately every two years, outpacing the slowing of Moore's Law.
But the most consequential decision of this era wasn't a product launch. It was a quiet architectural choice: in 2003, NVIDIA added IEEE-compatible 32-bit floating point (FP32) to its shader processors. This single change meant that scientific code written for CPUs could, in principle, run on NVIDIA GPUs. It was the first step toward CUDA.
The CUDA Bet (2006–2012)
In November 2006, NVIDIA launched CUDA (Compute Unified Device Architecture) alongside the GeForce 8800 GTX. CUDA allowed developers to write general-purpose code in C that would run on NVIDIA GPUs — not just graphics, but physics simulations, financial modeling, molecular dynamics, anything that could be parallelized.
It was the most important decision in NVIDIA's history. And it nearly destroyed the company.
The problem: CUDA added significant cost to every GeForce GPU, because it required additional transistors, memory, and software infrastructure. Gamers didn't know CUDA existed and wouldn't pay more for it. But Jensen Huang insisted on putting CUDA on every GeForce card — even the cheapest consumer models.
"Install base defines an architecture. Not... Everything else is secondary."
Jensen Huang on the Lex Fridman Podcast, March 2026(1)
The strategy was simple but brutal: subsidize the platform to build the install base. GeForce was already selling millions of units per year. By putting CUDA on every card, NVIDIA would put a parallel supercomputer in the hands of every researcher, every student, every scientist with a gaming PC.
The cost was enormous. NVIDIA was a 35% gross margin company, and CUDA increased GPU costs by roughly 50%. The company's market capitalization dropped from approximately $8 billion to $1.5 billion. It took nearly a decade for the bet to pay off.
Meanwhile, Jensen went to universities. NVIDIA wrote textbooks, taught classes, gave away development kits, and funded GPU computing research. The CUDA ecosystem grew slowly — from a few hundred researchers in 2007 to thousands by 2010.
The early CUDA applications were in scientific computing: weather simulation, molecular dynamics, astrophysics, computational finance. These fields had the same computational pattern that games did — massive parallelism — but the audience was researchers, not gamers.
What nobody anticipated was that the same pattern would define the next revolution in computing.
The AlexNet Moment (2012)
In September 2012, three researchers — Geoffrey Hinton, Ilya Sutskever, and Alex Krizhevsky — entered the ImageNet Large Scale Visual Recognition Challenge with a neural network called AlexNet. They trained it on two NVIDIA GeForce GTX 580 gaming GPUs — consumer hardware that cost $500 each.
AlexNet obliterated the competition, jumping image classification accuracy from 74.3% to 84.7%. Twenty years of hand-engineered computer vision features were rendered obsolete in a single presentation.
The recipe was absurdly simple: bigger neural networks + more data + GPU compute = dramatically better AI.
Google acquired Hinton's team for $44 million within months. The deep learning revolution had begun, and NVIDIA GPUs were its engine. The same CUDA platform that Jensen Huang had been subsidizing for six years suddenly became the most important software stack in AI.
"The recipe — more GPUs, more data, bigger models — became the blueprint for GPT-1 through GPT-5 and is the reason Sutskever co-founded OpenAI in 2015."
Every major AI lab — Google Brain, Facebook AI Research, Baidu, DeepMind — started buying NVIDIA GPUs in bulk. The data center business, which barely existed in 2012, began its exponential climb.
From Gaming to AI Infrastructure (2012–2020)
The years following AlexNet saw NVIDIA systematically pivot from a gaming company to an AI infrastructure company — while never abandoning gaming.
| Year | GPU Architecture | AI Milestone |
|---|---|---|
| 2012 | Kepler | AlexNet trained on GeForce GTX 580s |
| 2014 | Maxwell | Energy efficiency breakthrough, cuDNN library |
| 2016 | Pascal (P100) | First GPU designed for deep learning, NVLink |
| 2017 | Volta (V100) | Tensor Cores for mixed-precision training |
| 2018 | Turing | RT cores + Tensor Cores, first ray-tracing GPU |
| 2020 | Ampere (A100) | 3rd-gen Tensor Cores, multi-instance GPU |
Key milestones of this era:
2016: The DGX-1. NVIDIA shipped the world's first purpose-built deep learning supercomputer directly to OpenAI. Jensen Huang personally delivered it to Sam Altman. The DGX-1 packed eight Tesla P100 GPUs into a single box — the equivalent of 250 conventional servers. It was NVIDIA's declaration that AI infrastructure was now a product category.
2017: The Transformer Paper. Google researchers published "Attention Is All You Need," introducing the transformer architecture that would power GPT, BERT, Claude, Gemini, and virtually every modern AI model. Transformers were massively parallelizable — perfectly suited to GPU computation. NVIDIA's hardware advantage compounded.
2018: Jensen's GTC Keynote Prediction. At the GPU Technology Conference, Huang predicted that AI would become "the most powerful technology force of our time" and that NVIDIA would be its primary infrastructure provider. Analysts were skeptical — gaming still accounted for the majority of NVIDIA's revenue. They would not be skeptical for long.
2020: The Mellanox Acquisition. NVIDIA acquired Mellanox Technologies for $6.9 billion — its largest acquisition ever. Mellanox made InfiniBand networking equipment, the high-speed fabric connecting GPUs in data centers. The acquisition signaled NVIDIA's shift from chip company to systems company. You couldn't just build the best GPU anymore; you had to build the best rack.
2020: The Failed Arm Acquisition. NVIDIA announced a $40 billion deal to acquire Arm Holdings from SoftBank. If completed, it would have given NVIDIA control over the instruction set architecture used in virtually every smartphone and a growing number of data center chips. Regulators in the US, EU, UK, and China blocked the deal, and it collapsed in February 2022. NVIDIA designed its own ARM-based CPU (Grace) instead.
(update) The ChatGPT Explosion (2022–2023)
Everything changed on November 30, 2022, when OpenAI released ChatGPT. The chatbot reached 100 million users in two months — the fastest-growing consumer application in history. And every single token it generated ran on NVIDIA GPUs.
The demand shock was instantaneous. AI labs, hyperscale cloud providers (Microsoft, Google, Amazon, Oracle), and enterprises raced to acquire NVIDIA's latest GPUs. The H100 (Hopper architecture, launched March 2023) became the most sought-after piece of hardware in the world. Wait times stretched to 6-12 months. A single H100 that listed for $25,000-$40,000 sold for over $50,000 on secondary markets.
NVIDIA's financial transformation was staggering:
| Fiscal Year | Revenue | Data Center Revenue | Market Cap (Year-End) |
|---|---|---|---|
| FY2023 (Jan 2023) | $27.0B | $15.0B | ~$360B |
| FY2024 (Jan 2024) | $60.9B | $47.5B | ~$1.5T |
| FY2025 (Jan 2026) | $130.5B | $115.2B | ~$3.4T |
In June 2023, NVIDIA crossed $1 trillion in market capitalization — the first semiconductor company to do so. In February 2024, it crossed $2 trillion. In June 2024, it briefly surpassed both Apple and Microsoft to become the most valuable company in the world at over $3 trillion.
Jensen Huang went from a respected but niche tech CEO to a global figure. He appeared on the cover of TIME magazine, received standing ovations at conferences around the world, and became known for his signature black leather jacket — which he has worn at every major public appearance for decades.
(update) Blackwell, Grace Blackwell, and Rack-Scale Computing (2024–2025)
The Hopper era (H100/H200) was just the beginning. In March 2024, NVIDIA unveiled the Blackwell architecture at GTC 2024 — the largest chip NVIDIA had ever built, with 208 billion transistors.
The Blackwell B200 GPU was designed for one thing: training and running the largest AI models in the world. But the real innovation wasn't the chip alone — it was the Grace Blackwell NVL72 rack, which connected 72 Blackwell GPUs via NVLink 5 into a single computing domain.
This was Jensen's vision of extreme co-design in action: GPU, CPU (Grace), memory (HBM3e), networking (NVLink, InfiniBand), and power/cooling designed as a single system, not discrete components bolted together.
"The problem no longer fits inside one computer to be accelerated by one GPU. You want 10,000 computers to go a million times faster... We just gotta bring every technology to bear."
Jensen Huang, Lex Fridman Podcast #494(1)
The B200 shipped to customers in late 2024, with Microsoft, Google, Amazon, Meta, Oracle, and xAI among the first buyers. Elon Musk's xAI built Colossus — a 200,000-GPU supercomputer in Memphis, Tennessee — in just four months using NVIDIA hardware. Jensen publicly praised Musk's speed-of-execution approach.
NVIDIA GTC 2024 Keynote — Jensen Huang unveils the Blackwell architecture, the 208-billion-transistor B200 GPU, and the GB200 NVL72 rack-scale system.
Key 2024–2025 products and announcements:
| Date | Product / Event | Significance |
|---|---|---|
| Mar 2024 | Blackwell B200, GB200 NVL72 | 208B transistors, rack-scale AI computing |
| Mar 2024 | NVIDIA NIM | Inference microservices for enterprise AI deployment |
| Jun 2024 | $3T market cap | Briefly most valuable company in the world |
| Sep 2024 | H200 shipments ramp | 141 GB HBM3e memory, 2x inference vs H100 |
| Nov 2024 | Blackwell customer shipments begin | Record demand across cloud providers |
| Jan 2025 | NVIDIA Cosmos | World model for robotics and autonomous vehicles |
| Mar 2025 | DGX Spark, DGX Station | Desktop AI supercomputers for researchers |
(update) GTC 2026: Vera Rubin and the Agentic Era (2026)
At GTC 2026 (March 2026), Jensen Huang delivered what many called his most consequential keynote — not because of a single product, but because it redefined what NVIDIA builds and why.
The headline announcement was Vera Rubin, NVIDIA's next-generation AI computing platform named after the astronomer who discovered evidence for dark matter. Vera Rubin includes:
- A new GPU architecture with enhanced Tensor Cores optimized for agentic workloads
- The Vera CPU — NVIDIA's most powerful ARM-based processor
- NVLink 72 — connecting 72 GPUs as a single computing domain
- Storage accelerators — a new component class for agent tool-use and file access
- 1.3 million components per rack from 200 suppliers, shipped as pre-assembled supercomputers weighing 2–3 tons each
The architectural shift from Grace Blackwell to Vera Rubin was telling. Grace Blackwell was optimized for running mixture-of-experts large language models. Vera Rubin was optimized for agentic AI — systems that don't just generate text but use tools, access files, spawn sub-agents, and take autonomous action.
Jensen also announced:
- NemoClaw — an enterprise-grade security framework layered on top of OpenClaw, running in NVIDIA's open-source OpenShell sandbox runtime. It wraps OpenClaw instances with policy-based YAML guardrails, model constraints for safety validation, and local-first compute on NVIDIA hardware — designed to bring the open agentic ecosystem securely to the enterprise (the open install base below; the monetized enterprise layer above)
- Nemotron — NVIDIA's open-source AI models, with a coalition including Black Forest Labs, Perplexity, Mistral, and Cursor
- CUDA 13.2 — continuing the evolution of the computing platform that started it all in 2006
- Grok integration — partnerships expanding NVIDIA's agentic AI ecosystem
The same week, Jensen appeared on the Lex Fridman Podcast (#494), the All-In Podcast, and gave a Stratechery interview — a media blitz that shaped the industry's understanding of where AI infrastructure is heading.
Jensen Huang on the Lex Fridman Podcast #494 (March 2026) — CUDA origin story, four scaling laws, extreme co-design, 60+ direct reports, and why "inference is thinking."
His most viral quote from the All-In Podcast: "If you're a $500K developer, you should be spending $250K on AI tokens." The statement — that companies should invest as much in AI compute as they do in human engineers — captured the economics of the agentic era in a single sentence.
(update) May 2026: SCSP "Memos to the President" — Six Months of Agent Breakthroughs
SCSP "Memos to the President" Episode 43 (May 2026) — Jensen Huang on generative computing, the five-layer AI cake, reindustrialization, physical AI, agent harnesses, and AI export policy. The canonical 2026 Jensen interview.
In Jensen Huang's second SCSP interview with Ylli Bajraktari (May 2026, Crystal City), he reframed how the industry moved from large language models to fully agentic systems in just six months:
"The big breakthrough from large language models to chatbots was reinforcement learning human feedback. The big breakthrough from large language models to agentic systems is a system called harnesses."
Jensen Huang, SCSP "Memos to the President" (May 2026)
The pre-training has improved, reinforcement learning has improved, and reasoning has improved. But the bigger leap was harnessing large language models so they can connect to ground truth, do research, drive a web browser, reason with persistent memory, and communicate with other agents. Jensen praised both OpenAI Codex and Claude Code as examples of the harness pattern done well: the vast majority of software tasks are now automated end-to-end, not because the underlying model became smarter, but because the harness around it became real.
This matters for NVIDIA because every harness call multiplies token throughput. A single user prompt that used to generate 500 tokens of chat output now triggers minute-long reasoning chains, browser sessions, file reads, sub-agent dispatches, and tool invocations — often tens of thousands of tokens of generative work behind one human prompt. Vera Rubin was designed for exactly this workload pattern.
The Radiologist & Software Engineer Paradox
Jensen's most pointed argument from the SCSP interview is that AI is creating jobs, not destroying them — and that scaring young people away from careers AI is "supposed" to wipe out is actively hurtful to society. He pulled two case studies:
| Prediction (2015) | Reality (2026) |
|---|---|
| "Radiologists will be the first profession AI eliminates." | Radiologists are in short supply. AI permeates every aspect of radiology, but the purpose of the job — diagnosing disease — requires more humans, not fewer. Reading scans is the task; diagnosis is the purpose. |
| "Software engineering jobs are gone forever." | Every company including NVIDIA is hiring more software engineers than ever. Coding is the task; the purpose is innovating, solving problems, and connecting unrelated ideas. AI handles the typing — humans handle the imagination. |
"The task of our job and the purpose of our job are related, not the same. If you applied that to me, you'd come to the conclusion that what Jensen does for a living is tap on phones and talk. And tapping on phones and talking, AI has done that just fine. Therefore, my job should be gone. But I'm busier than ever."
Jensen Huang, SCSP "Memos to the President" (May 2026)
The deeper point is economic: AI compresses the cost of writing a billion lines of code, but the demand for code is not capped at a billion lines. With AI, organizations can pursue a trillion lines of code worth of imagination — in healthcare, science, manufacturing, retail, robotics, and countless fields where typing was the bottleneck. AI has already created over half a million jobs in the last couple of years according to Jensen — and adoption is the multiplier.
🤖 Physical AI in May 2026: Robo-Taxis, Humanoid Robots, Alpamayo
In July 2024, Jensen predicted physical AI would be the next frontier. Ten months later, the prediction is materializing in production:
Robo-Taxis: The Science Is Solved
Self-driving robo-taxis are no longer a research problem. NVIDIA shipped Alpamayo — what Jensen calls "the world's first thinking robo-taxi" — software that can encounter a never-before-seen situation, decompose it into mundane known patterns ("I've seen this, I've seen that, I've seen that"), reason from composition, and take the right action. Reasoning systems have unlocked the long tail of edge cases that pure pattern matching couldn't.
"Robo-taxis are here. The science is now completely solved. And even the engineering is just around the corner."
Jensen Huang, SCSP "Memos to the President" (May 2026)
Humanoid Robots: Right Around the Corner
The remaining gap for humanoids is not the AI — it's the body. Generative video models can already prompt a video of a hand picking up a coffee cup and taking a sip. If you can generate the video, you can generate the robot action. The blockers are mechatronics:
- Material science — the body must be light enough to be safe (an 80-pound robot tipping over is recoverable; a 300-pound one is not) but strong enough to be useful
- Motors and hands — fine articulation that matches the AI's intent
- Battery technology — sustained operating time without thermal issues
- Sensors — proprioception, vision, touch
Once those engineering layers catch up to the AI, humanoid robots become consumer-scale products.
Tokens for the Physical World
The GTC 2026 keynote opening framed it perfectly: tokens are not just text. They harness clean energy, unlock the secrets of the stars, help robots learn in virtual worlds and perfect actions in the physical world, forge new paths, clear ground for the harvest, work where human hands cannot, and help the smallest hearts beat stronger. Every robot — autonomous vehicle, humanoid, factory arm, agricultural sensor — is a token consumer downstream of NVIDIA's AI factories.
🌏 China Policy & Export Controls (May 2026 Update)
Jensen has been the industry's most consistent advocate for treating AI export policy as dynamic rather than static. In the May 2026 SCSP interview, he updated his stance with new data:
- NVIDIA's market share inside China for advanced AI chips has dropped to effectively zero under current export controls
- Conceding a market the size of China to domestic Chinese chipmakers (Huawei, Cambricon, etc.) does not make long-term strategic sense
- The five-layer cake framework matters here: at the chip layer, NVIDIA leads — but ceding chip share weakens the funding base that compounds into the model and infrastructure layers
- At the model layer, the US is ahead but China is close behind, with extraordinary AI researcher density that Jensen calls "one of their national treasures"
- At the adoption layer, Jensen is "very concerned" — the rest of Asia embraces AI with enthusiasm while the US debates "cinematic, science-fiction" framings of risk
Jensen's policy proposal is balanced: provide the best and the most to American companies first, ensure the American tech stack wins worldwide through diffusion, and avoid policies whose first-order effect is conceding the world's largest single AI market. "Maybe it made sense at the time, but the policy needs to be dynamic and stay with the times."
🛡️ OpenShell: Sandboxing Open-Source Agents at Industrial Scale
A subtle but consequential announcement from the SCSP interview was the deeper architecture of OpenShell — NVIDIA's open-source sandbox runtime that wraps OpenClaw instances in a controlled execution environment.
The naming is intentional. OpenClaw is the open-source agent framework with the cuddly lobster mascot; OpenShell is the shell around the open claw — a virtual cage that lets enterprises adopt agentic AI without exposing sensitive data, networks, or tools. Three layers of control:
- Sandbox / virtual environment — agents execute in an isolated runtime, not on the host machine
- Policy engine — fine-grained controls over what information the agent can read, send out, or persist; PII redaction; topic boundaries
- Privacy controls — per-instance rules so a single deployment can serve users with different data-access scopes
Jensen's broader argument is asymmetric defense. The way to defend against a super-agent cybersecurity attack is not another super-agent. It is a massive swarm of open-source defensive agents, trained for specific defense tasks, coordinated together. CrowdStrike, Palo Alto Networks, Cisco, Microsoft — all rely on open-source primitives to build that swarm. OpenShell makes the swarm safe.
"You're not going to defend against a super agent with another super agent. You're going to defend it with massive swarms. To use asymmetry to your advantage requires open-source technology."
Jensen Huang, SCSP "Memos to the President" (May 2026)
This is also a strategic play: the same OpenClaw + OpenShell stack that the cybersecurity industry runs on is the foundation of NemoClaw, NVIDIA's enterprise-grade fork. Open install base below, monetized enterprise layer above — the CUDA playbook applied to agents.
📋 NVIDIA's GPU Architecture Timeline
Every major NVIDIA GPU architecture maps directly to an AI capability breakthrough — from CUDA's debut on Tesla in 2006 through tensor cores on Volta, transformer-scale training on Ampere and Hopper, and the 208 billion transistor Blackwell chip that powers today's frontier model training. This timeline shows how each generation expanded what AI could practically do.
| Year | Architecture | Key GPU | Transistors | AI Significance |
|---|---|---|---|---|
| 2006 | Tesla (G80) | GeForce 8800 GTX | 681M | First CUDA GPU, general-purpose computing |
| 2010 | Fermi | Tesla M2050 | 3B | First GPU with ECC memory for scientific computing |
| 2012 | Kepler | GeForce GTX 680 | 3.5B | AlexNet trained on GTX 580 (prior gen) |
| 2014 | Maxwell | GeForce GTX 980 | 5.2B | cuDNN library, energy efficiency leap |
| 2016 | Pascal | Tesla P100 | 15.3B | First NVLink, mixed-precision hints |
| 2017 | Volta | Tesla V100 | 21.1B | First Tensor Cores, dedicated AI silicon |
| 2018 | Turing | GeForce RTX 2080 | 18.6B | Ray tracing + Tensor Cores in consumer GPUs |
| 2020 | Ampere | A100 | 54.2B | 3rd-gen Tensor Cores, multi-instance GPU |
| 2022 | Hopper | H100 | 80B | 4th-gen Tensor Cores, transformer engine |
| 2024 | Blackwell | B200 | 208B | Rack-scale computing, NVLink 72 |
| 2026 | Vera Rubin | TBA | TBA | Agentic AI, storage accelerators |
NVIDIA's pace of improvement: in the last 10 years, Moore's Law would have improved computing about 100x. Through extreme co-design, NVIDIA improved AI computing by over 1,000,000x — a million-fold improvement that Jensen Huang cites as proof that accelerated computing has fundamentally decoupled from transistor scaling.
💰 The Financial Story: From $42 Million IPO to $3.4 Trillion
NVIDIA's market capitalization trajectory is one of the most dramatic in corporate history:
| Date | Market Cap | Catalyst |
|---|---|---|
| Jan 1999 | $600M | IPO on Nasdaq |
| 2002 | $8B | GeForce dominance, Xbox GPU |
| 2008 | $1.5B | Post-CUDA cost impact + financial crisis |
| 2016 | $30B | Data center GPU demand begins |
| 2020 | $300B | COVID-19 remote work + gaming boom |
| Jun 2023 | $1T | First semiconductor company at $1T |
| Feb 2024 | $2T | AI infrastructure demand explosion |
| Jun 2024 | $3T | Briefly most valuable company |
| Mar 2026 | $3.4T+ | Sustained AI infrastructure buildout |
Revenue growth over the past five years:
| Fiscal Year (ending Jan) | Revenue | YoY Growth | Data Center % |
|---|---|---|---|
| FY2021 | $16.7B | +53% | 40% |
| FY2022 | $26.9B | +61% | 44% |
| FY2023 | $27.0B | +0.2% | 56% |
| FY2024 | $60.9B | +126% | 78% |
| FY2025 | $130.5B | +114% | 88% |
The data center business has gone from 40% to 88% of total revenue in five years. Gaming — the business that built the company — now accounts for less than 10% of revenue. The transformation is complete.
NVIDIA's gross margins exceed 70% — remarkable for a hardware company and a testament to the software moat (CUDA ecosystem) that prevents customers from switching to competitors even as AMD, Intel, Google (TPU), Amazon (Trainium), and others invest billions in alternatives.
🔎 Key Partnerships and the NVIDIA Ecosystem
NVIDIA's power extends far beyond its own chips. The company has built an ecosystem that touches every major AI company in the world:
Hyperscale Cloud Providers
- Microsoft Azure — largest cloud GPU customer, powers OpenAI's infrastructure
- Google Cloud — runs NVIDIA GPUs alongside Google's own TPUs
- Amazon AWS — offers NVIDIA GPU instances (P5, P4d) alongside its Trainium chips
- Oracle Cloud — invested heavily in NVIDIA GPU capacity (Stargate Project partner)
- Meta — built massive NVIDIA GPU clusters for Llama model training
AI Labs
- OpenAI — GPT-5, o3, and all models trained on NVIDIA hardware
- Anthropic — Claude models trained on NVIDIA GPUs (NVIDIA invested in Anthropic's Series C)
- Google DeepMind — Gemini models use both TPUs and NVIDIA GPUs
- xAI — Grok models trained on 200,000 NVIDIA GPUs (Colossus)
- Meta AI — Llama open-source models trained on NVIDIA infrastructure
Enterprise Software
- NVIDIA AI Enterprise — full-stack software for deploying AI in production
- NeMo — framework for building custom large language models
- NIM (NVIDIA Inference Microservices) — optimized inference containers
- Omniverse — platform for industrial digital twins and simulation
- DRIVE — autonomous vehicle computing platform
Open-Source Ecosystem
- Nemotron — NVIDIA's open-source AI models (March 2026 coalition with Perplexity, Mistral, Cursor, Black Forest Labs)
- NemoClaw — open security framework for agentic systems
- CUDA libraries — cuDNN, cuBLAS, TensorRT, Triton — the software that makes AI frameworks fast
🤯 Jensen Huang: The CEO Who Never Left
Jensen Huang has been CEO of NVIDIA for over 32 years — from the Denny's founding in 1993 to a $3.4 trillion company in 2026. No other tech CEO has led a single company through such a dramatic transformation over such a long period.
Management Philosophy: 60 Direct Reports, No 1-on-1s
Jensen's organizational philosophy is as unconventional as his technical decisions. He has 60+ direct reports — almost all with deep engineering expertise. He does not hold scheduled one-on-one meetings.
"We present a problem and all of us attack it... because we're doing extreme co-design. And literally, the company is doing extreme co-design all the time."
Jensen Huang, Lex Fridman Podcast #494(1)
When someone presents a problem — say, a cooling challenge in the Vera Rubin rack — the memory expert, the networking expert, the power delivery expert, and the GPU architect are all in the room, listening. Anyone can contribute. Anyone who should have contributed but didn't gets called out.
The organizational structure mirrors the product philosophy: just as NVIDIA co-designs across the full technology stack, the company co-designs across the full organizational stack. No information silos, no privileged access.
Belief-Shaping as Leadership
Jensen doesn't lead through annual manifestos or big organizational restructurings. He shapes belief systems gradually, in public, step by step.
When he decided NVIDIA should acquire Mellanox (networking), he had been discussing networking challenges with his team for months. By the announcement day, the response was: "What took you so long?"
He applies the same approach externally. GTC keynotes plant ideas two to three years before products ship. When Vera Rubin launched at GTC 2026, Jensen had been describing the agentic architecture schematic for over two years. The industry had been primed.
"By the time that I announce something, everybody's saying, 'You know, what took you so long?'"
Jensen Huang, Lex Fridman Podcast #494(1)
Personal Details
- Net worth: Exceeds $120 billion (March 2026), making him one of the 15 wealthiest people in the world
- NVIDIA tattoo: Jensen famously got the NVIDIA logo tattooed on his arm after the company's stock hit a certain milestone
- Leather jacket: His signature black leather jacket has become an icon in the tech industry — he wears it at every public appearance
- "Speed of light" methodology: A 30-year practice of comparing all work against physical limits — asking teams not just "how fast?" but "how close to the theoretical maximum?"
- Family: Married to Lori Huang. Two children. Lori is a descendant of a co-founder of Super Micro Computer.
🤔 What Makes NVIDIA Different?
The CUDA Moat
NVIDIA's dominance isn't just hardware — it's the software ecosystem. CUDA has 5+ million developers, 3,000+ GPU-accelerated applications, and deep integration into every major AI framework: PyTorch, TensorFlow, JAX, and Hugging Face.
Switching from NVIDIA to a competitor means rewriting software at every layer — from low-level kernels to framework integrations to application code. AMD's ROCm and Intel's oneAPI are technically capable alternatives, but the ecosystem gap remains enormous. Google's TPUs are competitive for internal Google workloads but lack the universal developer platform.
This is exactly the lesson Jensen learned from the x86 architecture: install base defines architecture. The most elegant hardware in the world loses to the platform with the most developers.
NVIDIA vs. the Competition (2026)
| NVIDIA | AMD | Intel | Google TPU | Amazon Trainium | |
|---|---|---|---|---|---|
| AI training share | 90%+ | ~5% | <2% | ~3% (internal) | <1% |
| Software ecosystem | CUDA (5M+ devs) | ROCm (growing) | oneAPI (early) | JAX/XLA (Google) | Neuron SDK |
| Latest AI chip | Blackwell B200 | MI300X | Gaudi 3 | TPU v5p | Trainium2 |
| Transistors | 208B | 153B | N/A | N/A | N/A |
| Rack-scale system | NVL72 (72 GPUs) | MI300X cluster | Gaudi 3 cluster | TPU pods | UltraCluster |
| Custom CPU | Grace (ARM) | No (uses AMD EPYC) | Xeon | No | Graviton |
| Networking | NVLink + InfiniBand | Infinity Fabric | Gaudi NIC | ICI | EFA |
| Open-source models | Nemotron | No | No | Gemma | No |
NVIDIA's moat is not one technology — it's the full stack. Competitors match one layer but can't match all seven simultaneously.
Extreme Co-Design
NVIDIA is no longer a chip company. It's a systems company that designs GPU, CPU, memory interfaces, networking, switching, power delivery, cooling, software, and the rack itself as a single optimized system.
Jensen calls this extreme co-design, and it's the reason NVIDIA's performance improvements outpace what transistor scaling alone would deliver. Moore's Law has improved computing roughly 100x in the past decade. NVIDIA's extreme co-design has improved AI computing by over 1,000,000x.
The Four Scaling Laws
At GTC 2026 and on the Lex Fridman Podcast, Jensen articulated four scaling laws that define the future of AI:
- Pre-training scaling — larger models + more data = smarter AI. Data increasingly synthetic, no longer limited by human-generated text.
- Post-training scaling — fine-tuning, RLHF, and reinforcement learning continue to scale with synthetic data.
- Test-time scaling — reasoning, planning, and search at inference time. Jensen's key insight: "Inference is thinking, and thinking is way harder than reading."
- Agentic scaling — spawning sub-agents, using tools, doing research. "It's so much easier to scale NVIDIA by hiring more employees than it is to scale myself."
These four laws form a reinforcing loop. Agentic systems create new experiences; good experiences get memorized (pre-training), refined (post-training), enhanced at test time, and deployed via agents. The cycle accelerates. And it all requires more compute.
"Intelligence is gonna scale by one thing, and that's compute."
Jensen Huang, Lex Fridman Podcast #494(1)
Token Economics
NVIDIA's hardware prices keep going up — a Vera Rubin rack costs significantly more than a Blackwell rack. But the cost per AI token keeps going down by roughly an order of magnitude per year. This is the virtuous cycle: better hardware → cheaper tokens → more AI usage → more demand for hardware.
Jensen's framing: "Our computer price is going up, but our token generation effectiveness is going up so much faster that token cost is coming down."
🛡️ NemoClaw and the Battle for Agentic Engineering
NemoClaw is more than a security framework — it represents a fundamentally different bet on how enterprises will adopt agentic AI. While OpenAI and Anthropic spent 2025 learning that most companies lacked the expertise to deploy tools like Codex and Claude Code in production, NVIDIA walked onto the stage and said: you developers can figure this out.
NVIDIA open-sourced what OpenAI and Anthropic want you to pay consultants for — NemoClaw, Rob Pike's timeless engineering rules, and why simple scales better than complex in the agentic era.
The contrast is stark. OpenAI and Anthropic are publicly tying up with major consulting firms — they recognized that their solutions were too complicated for engineering teams at enterprises to successfully adopt. Teams weren't seeing the same speed-ups that Anthropic and OpenAI achieved internally. Jensen's answer: an open-source framework that wraps OpenClaw in enterprise-grade compliance and security, runs on Linux, and trusts developers to build.
How NemoClaw Works
NemoClaw is an add-on to OpenClaw, not a replacement. It runs inside OpenShell, NVIDIA's open-source sandbox runtime, which wraps OpenClaw instances with three layers of control:
- Policy-based guardrails — YAML declarations that agents must follow, defining boundaries for what actions are permitted
- Model constraints — safety validation that also ensures NVIDIA serves the model (a strategic move from chips to the agentic value chain)
- Local-first compute — designed to run on NVIDIA hardware locally, keeping sensitive enterprise data off the public internet
The strategic play is clear: Jensen is building an ecosystem where the hundreds of thousands of developers already using OpenClaw indirectly contribute value to NemoClaw, which NVIDIA sells to the enterprise. It's the CUDA playbook applied to agents — build the open install base, then monetize the enterprise layer.
Rob Pike's Five Rules, Rediscovered for Agents
What makes NVIDIA's approach work isn't just corporate strategy — it's that their engineers understand timeless software principles. Rob Pike, co-creator of Unix and Go, wrote five rules of programming decades ago. Every one of them applies directly to agentic engineering:
| Pike's Rule | Original Meaning | Agentic Application |
|---|---|---|
| 1. You can't tell where a program spends its time | Don't add speed hacks until you've proven where the bottleneck is | Agentic systems have surprising bottlenecks — don't optimize prematurely |
| 2. Measure before you tune | Don't tune for speed until you've measured and one part overwhelms the rest | Baseline your LLM responses and agent performance before making changes |
| 3. Fancy algorithms are slow when N is small | Simple scales well; don't get fancy until proven necessary at scale | Simple agentic architectures scale better than complex ones — LLMs abstract edge-case complexity |
| 4. Fancy algorithms are buggier | Use simple algorithms for simple data structures | Complex agentic systems are extremely hard to debug — simplify ruthlessly |
| 5. Data dominates | Right data structures make algorithms self-evident | Data engineering is the key to good agentic systems — fix your data, and agent behavior becomes self-evident |
Rule five is the most important for the agentic era. If you've organized your data well — linter configs, documented builds, dev containers, an agents.md file — agent behavior becomes self-evident. It's Rob Pike's insight from the 1980s, rediscovered by every team that ships production agents in 2026.
Five Hard Problems in Production Agent Deployment
The production challenges teams face with agents aren't new computer science — they're old engineering problems wearing new clothes:
1. Context compression. Long-running agent sessions fill up context windows, even million-token ones. Every compression strategy is lossy. The best approach (incremental summarization with structured sections for intent, file modifications, decisions, and next steps) outperforms both OpenAI's opaque compact endpoint and Anthropic's full-summary regeneration. The mitigation: think in milestones, and use multi-agent frameworks that let agents complete chunks of work and hand off context cleanly.
2. Codebase instrumentation. This is Pike's Rule #2 — measure. Making a codebase agent-ready means being able to baseline performance: what does latency look like, what does a good set of responses look like, and do you have a golden test set? Decades-old discipline, but critical when you're giving autonomous agents real power.
3. Strict linting. Pike's Rules #3 and #4 — simplicity reduces bugs. Production agent code needs extremely strict linting because agents are, by definition, lazy developers happy to throw work off their plates. Without a strict linter enforcing simplicity, agent-generated code drifts toward complexity that humans can't maintain.
4. Multi-agent coordination. The industry is converging on planner-executor patterns for long-running multi-agent work. Don't over-complicate it. Build the simplest possible agentic pipeline first — you can always add complexity if measurement proves you need it.
5. Specification fatigue. The hardest problem. Teams struggle to define specs clearly upfront. If you give an agent a context window, you must ensure the context graph is clean — navigable hierarchies, not stuffed prompts. Humans have to be less lazy if they want agents to do good work. That's never how it's been sold, but it's always been true of good engineering.
The lesson from NemoClaw — and from every team successfully deploying agents in production — is that the chaos is not new. Consultants profit from presenting it as new. But if we anchor in the engineering principles we've always known and walk forward from there, the path to productive agentic systems is simpler than the hype suggests. NVIDIA's engineers, close to the kernel and the metal, understand this better than most.
🧰 The Harness Layer: Where 2026 Value Concentrates
In Jensen's May 2026 SCSP interview and Palantir CTO Shyam Sankar's April 2026 a16z interview, two of the most consequential CTOs in AI infrastructure independently named the same primitive: the harness. A harness is the orchestration software around a model that gives it ground truth, memory, browsers, tool use, and sub-agent dispatch. Models are getting commoditized; harnesses are where alpha accrues.
| Reference Harness | Specialty | Sandbox Model | Memory | Tools | Best For |
|---|---|---|---|---|---|
| OpenAI Codex | SWE | Hosted | Per-repo session | Built-in dev tools | Production code automation |
| Anthropic Claude Code | SWE | Local + hosted | Project memory | Built-in dev + MCP | Engineer-pair workflows |
| NVIDIA OpenShell + OpenClaw | Sandboxed enterprise agents | OpenShell virtual env | Pluggable (Milvus, Neo4j) | 22+ extensible | Enterprise data + compliance |
| MCP-native harnesses | Cross-tool interop | Per-server | Per-tool | Unbounded | Polyglot agent stacks |
| Taskade Agents v2 + Genesis | Workspace knowledge work | Workspace boundary | Workspace DNA | 22+ built-in + 100+ integrations | Teams, no-code builders, end users |
Jensen's framing is that harnesses are why the vast majority of software tasks are now automated end-to-end — not because models got smarter, but because the wrap around them got real. Sankar's framing is that as models commoditize, value flows to whoever owns the harness layer. Both are right.
🏗️ Who Builds Layer 5 of the AI Cake? (The Application Layer Question)
NVIDIA's five-layer AI cake (covered earlier in this post) names the layer Jensen is most worried about — the application and adoption layer — and is honest about not building it. NVIDIA builds chips, infrastructure, models, and harness primitives. It does not build the workspace where end users encode their company's ontology, attach their data, define their agents, and ship apps to their customers.
That is the question every CEO in May 2026 should be asking: for layer 5 of the AI cake, what do we put on the desk of every employee — and on the customer-facing surface — that captures the application-layer value?
Taskade Genesis is the workspace harness for layer 5. It sits directly above the model + harness primitives Jensen described and turns them into the surface where teams actually create value:
| Layer-5 Capability Jensen Describes | Taskade Genesis Implementation |
|---|---|
| Generative computing for everyone | One-prompt app generation; every output is novel and personalized |
| Agents with ground-truth access | Workspace-scoped Memory: Projects, databases, files, knowledge bases |
| Tool use + browsers | 22+ built-in agent tools, 100+ bidirectional integrations |
| Sub-agent dispatch | Multi-agent collaboration with planner/executor patterns |
| Persistent memory | Workspace DNA — agents remember across sessions; EVE stores its own memory as Taskade Projects |
| Sandbox / safe execution | Workspace boundary + 7-tier RBAC (Owner → Viewer) |
| Reindustrialization (knowledge work) | One-person companies with AI-empowered workflows; team output without team headcount |
| Open ecosystem | 11+ frontier models from OpenAI, Anthropic, Google — model-agnostic by design |
| Production deployment | Custom domains, password protection, Community Gallery publishing |
Where NVIDIA's reindustrialization thesis is about the physical-AI factory floor, Taskade Genesis is the same shift applied to the knowledge-work factory floor. The slingshot Sankar described — domain experts amplified 50–100× — runs through a workspace harness, not a chip.
🎯 The CEO Directive: What to Do This Week
If you read one section, read this one. Jensen's May 2026 SCSP interview and Sankar's April 2026 a16z interview converge on the same checklist for any CEO, founder, or leader trying to convert the 2026 paradigm shift into actual outcomes — not headlines.
┌─── THE 2026 LEADERSHIP CHECKLIST ────────────────────────────┐
│ │
│ 1. Pick your layer of the AI cake. Don't try to win all │
│ five. Pick where you can genuinely create alpha. │
│ │
│ 2. Stop telling young people their job is going away. │
│ Hire more software engineers, not fewer. Hire more │
│ radiologists. The task is automated; the purpose is │
│ expanding. │
│ │
│ 3. Pick a harness. Codex / Claude Code for SWE. │
│ Taskade Genesis for workspace + knowledge work. │
│ OpenShell + NemoClaw if you need an enterprise │
│ sandbox at scale. │
│ │
│ 4. Co-locate R&D and production. Innovation is downstream │
│ of productivity. If your team builds, ships, and │
│ operates inside the same workspace, the next │
│ innovation appears naturally. │
│ │
│ 5. Build alpha software, kill beta software. Audit your │
│ SaaS stack: which tools express your edge vs. make │
│ you generic? Cancel the second list. Re-build it as │
│ AI-native apps in your own workspace. │
│ │
│ 6. Treat AI compute like payroll. If you have a $500K │
│ developer, plan to spend $250K on tokens. The ROI is │
│ in the harness, not the model API. │
│ │
└──────────────────────────────────────────────────────────────┘
The companies that win the next phase will not be the ones with the largest model API contract. They will be the ones whose workspaces — Memory, Intelligence, Execution — are tuned tightest to their actual customers, data, and goals. NVIDIA gives you the infrastructure. Taskade Genesis gives you the workspace harness on top. Start your build →
⚡️ How NVIDIA Powers the AI You Use Every Day
Every AI product you interact with runs on NVIDIA infrastructure:
| AI Product | Company | NVIDIA Connection |
|---|---|---|
| ChatGPT and frontier models | OpenAI | Trained and served on NVIDIA GPUs in Microsoft Azure |
| Claude, Claude Code | Anthropic | Trained on NVIDIA GPUs; NVIDIA is an investor |
| Gemini | Uses NVIDIA GPUs alongside Google TPUs | |
| Grok | xAI | 200,000 NVIDIA GPUs in Colossus cluster |
| Llama | Meta | Trained on massive NVIDIA GPU clusters |
| Midjourney | Midjourney | Runs on NVIDIA GPU cloud infrastructure |
| Taskade Genesis | Taskade | Uses 11+ frontier models, all trained on NVIDIA hardware |
When you build an app with Taskade Genesis, you're using AI models that were trained on hundreds of thousands of NVIDIA GPUs over months of compute time. The agents, automations, and workspace intelligence in Taskade all depend on frontier models that exist because NVIDIA built the hardware to train them.
Beyond Chips: When AI Builds Your Tools
NVIDIA made AI training possible. Taskade Genesis makes AI application building accessible. Where NVIDIA provides the infrastructure layer — GPUs, networking, software — Taskade Genesis provides the workspace layer that turns AI capability into deployed, living applications.
Jensen's thought experiment about the "digital worker" — an AI that accesses files, does research, uses tools, and spawns sub-agents — maps directly to what Taskade builds:
| Jensen's Digital Worker | Taskade Feature |
|---|---|
| Access ground truth / file system | Projects, databases, file uploads |
| Research capability | AI agents with web browsing, 22+ built-in tools |
| Tool use | 100+ integrations, custom agent tools |
| External communication | Automations with Slack, email, webhooks |
| Sub-agent spawning | Multi-agent collaboration, automation branching |
The connection is Workspace DNA: Memory feeds Intelligence, Intelligence triggers Execution, Execution creates Memory — the same self-reinforcing loop that Jensen describes in his four scaling laws.
Try frontier AI models inside Taskade Genesis →
👉 How to Get Started with NVIDIA
Getting started with NVIDIA's AI ecosystem means choosing between hands-on GPU development (CUDA Toolkit, free at developer.nvidia.com) and managed platforms that abstract the hardware away. For developers, the CUDA Toolkit plus NGC pre-trained models is the fastest path; for teams that want AI without infrastructure, workspace platforms like Taskade Genesis run frontier models on NVIDIA hardware behind the scenes.
- CUDA Toolkit — free at developer.nvidia.com, includes compiler, libraries, debugging tools
- NVIDIA AI Enterprise — enterprise platform for deploying AI at scale
- DGX Cloud — cloud-based access to NVIDIA's most powerful GPU systems
- NGC Catalog — pre-trained models, Helm charts, and SDKs
- DGX Spark and DGX Station (announced GTC 2025) — desktop AI supercomputers for researchers and small teams
If you want to use AI without managing GPU infrastructure, platforms like Taskade abstract away the hardware layer entirely. You describe what you want, and AI agents — powered by frontier models trained on NVIDIA GPUs — build, deploy, and automate it.
💡 Pro Tip: You don't need to understand GPU architecture to use AI effectively. Taskade Genesis lets you build AI-powered applications with a single prompt — no CUDA required.
🚀 Quo Vadis, NVIDIA?
The next chapter of NVIDIA's story will be defined by three questions:
Can NVIDIA maintain its moat? AMD, Intel, Google (TPU), Amazon (Trainium), and a wave of AI chip startups are investing billions to challenge NVIDIA's dominance. But every year, the CUDA ecosystem grows deeper and the switching costs grow higher. Jensen's strategy of extreme co-design — optimizing across the full stack rather than just the chip — means competitors must match not just the silicon but the entire system.
How big is the AI market? Jensen has argued that the $1 trillion data center industry will be entirely rebuilt for accelerated computing — replacing CPU-based infrastructure with GPU-based infrastructure over the next decade. If he's right, NVIDIA's current revenue is early innings. If the AI buildout slows, the $3.4 trillion valuation faces a reckoning.
What happens when AI builds AI? Jensen's four scaling laws point to a world where AI agents create their own training data, spawn their own sub-agents, and improve recursively. NVIDIA's Vera Rubin platform is explicitly designed for this agentic future. The question is whether agentic AI creates demand that grows faster than NVIDIA can build — and whether the energy grid can keep up.
One thing Jensen Huang has never lacked is conviction. From the Denny's booth in 1993 to the most valuable company in the world, the thesis has been the same: parallel computing will eat everything. AI just proved him right.
🐑 Before you go... Looking for an AI tool that can seamlessly integrate with your task and project management workflow? Taskade AI has you covered!
💬 AI Chat: Got a question about your project? The AI Chat will give you all the answers and help your team stay on the same page. Decision-making has never been easier.
🤖 AI Agents: Tired of repetitive tasks? Let Taskade's AI Agents handle them. Agents work in the background, help your team save time, and let you focus on the big picture.
✏️ AI Assistant: Planning and organizing? You can use the AI Assistant to brainstorm ideas, generate content, organize tasks, and ensure you always know what to do next.
🔄 Workflow Generator: Starting a project can be a puzzle. Describe your project and the Workflow Generator will set things up for you, giving you a clear path to follow.
Want to give Taskade AI a try? Create a free account and start today! 👈
🔗 Resources
💬 Frequently Asked Questions About NVIDIA
Who is Jensen Huang?
Jensen Huang is the co-founder, president, and CEO of NVIDIA. Born in Tainan, Taiwan, in 1963, he moved to the United States as a child and earned degrees from Oregon State University and Stanford. He co-founded NVIDIA in 1993 at age 30 and has led the company for over 32 years — through the GPU invention, the CUDA bet, and the AI revolution. His net worth exceeds $120 billion as of March 2026.
What does NVIDIA stand for?
The name NVIDIA is derived from "invidia," the Latin word for envy. The founders chose it to suggest that their graphics technology would be the envy of the industry. The company is incorporated as NVIDIA Corporation and trades on the Nasdaq under the ticker NVDA.
When was NVIDIA founded?
NVIDIA was founded on January 25, 1993, by Jensen Huang, Chris Malachowsky, and Curtis Priem at a Denny's restaurant in San Jose, California. The company went public on the Nasdaq in January 1999.
What is the difference between NVIDIA and AMD?
Both NVIDIA and AMD make GPUs, but they differ significantly in AI. NVIDIA dominates AI training and inference with over 90% market share, powered by the CUDA software ecosystem with 5+ million developers. AMD competes in gaming GPUs and is growing its data center AI business with MI300X GPUs and the ROCm software stack, but its AI ecosystem is much smaller. NVIDIA also designs its own CPUs (Grace), networking (NVLink, InfiniBand via Mellanox), and full rack-scale systems — AMD primarily competes at the chip level.
What is NVIDIA worth?
As of March 2026, NVIDIA's market capitalization exceeds $3.4 trillion, making it the most valuable company in the world. The company generated $130.5 billion in revenue in fiscal year 2025 with gross margins above 70%.
Why is NVIDIA stock so expensive?
NVIDIA stock (NVDA) reflects the market's expectation that AI infrastructure spending will continue to grow exponentially. The company's revenue more than doubled in both FY2024 and FY2025, driven by data center GPU demand from AI labs, cloud providers, and enterprises. The CUDA software moat, 90%+ AI training market share, and expanding agentic AI workloads support the premium valuation.
What is Jensen Huang's net worth?
Jensen Huang's net worth exceeds $120 billion as of March 2026, derived almost entirely from his NVIDIA stock holdings (approximately 3.5% of outstanding shares). He is among the 15 wealthiest people in the world.
Does NVIDIA make CPUs?
Yes. NVIDIA designs the Grace CPU, an ARM-based processor optimized for AI and high-performance computing workloads. Grace is paired with NVIDIA GPUs in the Grace Blackwell and Vera Rubin platforms. NVIDIA also designed CPUs for the Tegra line (mobile/automotive). NVIDIA does not compete in the general-purpose x86 CPU market dominated by Intel and AMD.
What companies use NVIDIA GPUs for AI?
Virtually every major AI company uses NVIDIA GPUs. Key customers include Microsoft (Azure, OpenAI), Google Cloud, Amazon AWS, Meta, xAI, Oracle, Anthropic, Tesla, and hundreds of enterprise companies. Over 90% of AI model training worldwide runs on NVIDIA hardware.
What is NVIDIA DGX?
DGX is NVIDIA's line of purpose-built AI supercomputers. The DGX-1, shipped in 2016, was the first. The current lineup includes DGX B200 (data center), DGX Station (workgroup), and DGX Spark (desktop). Jensen Huang personally delivered the first DGX-1 to OpenAI in 2016.
Is NVIDIA a monopoly?
NVIDIA holds over 90% of the AI training GPU market, which has drawn scrutiny from regulators and competitors. However, competition exists from AMD (MI300X), Intel (Gaudi), Google (TPUs), Amazon (Trainium), and numerous AI chip startups. The US Department of Justice opened an antitrust investigation into NVIDIA in 2024 regarding its dominance in AI chips. NVIDIA argues its market position is the result of decades of R&D investment, not anticompetitive behavior.
What is NVIDIA's connection to gaming?
NVIDIA was founded as a gaming graphics company, and the GeForce brand remains iconic. However, gaming now accounts for less than 10% of NVIDIA's revenue (down from over 50% in 2020). The GeForce RTX series still leads the consumer GPU market, and gaming was essential to NVIDIA's strategy — it was the GeForce install base that carried CUDA to millions of researchers and developers, seeding the AI revolution.
How does NVIDIA relate to AI chatbots like ChatGPT and Claude?
Every major AI chatbot — ChatGPT, Claude, Gemini, Grok, and the AI models powering Taskade — was trained on NVIDIA GPUs. The training process requires thousands of GPUs running for weeks to months, processing trillions of tokens of text. NVIDIA doesn't build the AI models, but it builds the hardware and software (CUDA, cuDNN, TensorRT) that makes training them possible.
What did Jensen Huang say on the Lex Fridman Podcast?
On the Lex Fridman Podcast #494 (March 2026), Jensen Huang discussed NVIDIA's extreme co-design philosophy, the four AI scaling laws (pre-training, post-training, test-time, and agentic), the CUDA origin story (market cap dropping from $8B to $1.5B), his management approach (60+ direct reports, no 1-on-1s), agentic AI as "reinventing the computer," and why inference/thinking is "way harder than reading." He also explained NVIDIA's power grid waste proposal and the supply chain challenges of building 1.3-million-component racks.
What is NVIDIA NemoClaw?
NemoClaw is NVIDIA's enterprise security framework for agentic AI, announced at GTC 2026. It layers enterprise-grade compliance — policy-based YAML guardrails, model safety constraints, and local-first compute — on top of the open-source OpenClaw agent framework. NemoClaw runs inside NVIDIA's OpenShell runtime, allowing enterprises to deploy AI agents with controlled access to sensitive data, code execution, and external communication. It represents Jensen Huang's bet that developers can adopt agentic AI through open-source tooling rather than through expensive consulting engagements.
🧬 Build Your Own AI Applications
NVIDIA builds the infrastructure that trains AI models. Taskade Genesis lets you build applications on top of those models with a single prompt. Create custom AI agents, workflows, and automations — it's vibe coding that produces living software. Explore ready-made AI apps in the Community Gallery, or build your own.





