The Diner Napkin That Started a $3 Trillion Empire
It is 1993. A 30-year-old electrical engineer named Jensen Huang sits in a Denny's diner in Fremont, California โ a budget chain restaurant famous for its $6 breakfasts. Across from him are two friends: Chris Malachowsky and Curtis Priem. Between pancakes and coffee, the three of them sketch out an idea on whatever paper is available.
The idea: build a chip specifically designed to render computer graphics โ to make video games look real. Not a general-purpose chip like a CPU. A specialised one. They call the company Nvidia.
Nobody at the time thought this was a world-changing idea. It was a niche product for a niche market. But Jensen Huang had a particular way of thinking about problems โ one that would, thirty years later, make him one of the most powerful people in technology history.
Jensen Huang was born in Tainan, Taiwan in 1963. When he was nine, his parents โ wanting him to have a better education โ sent him to live with relatives in the United States. He ended up at a boarding school in Kentucky called Oneida Baptist Institute, sharing a dorm room with a roommate who, as Huang later recalled with a laugh, "had a tattoo and a knife."
He survived the culture shock, thrived academically, went to Oregon State University for his undergraduate degree in electrical engineering, and then got his Master's from Stanford. He worked at AMD and LSI Logic before the diner meeting in Fremont.
The detail that matters: at every step, Huang was obsessed not just with building things โ but with building systems. Not just the hardware, but the software to run it. Not just the chip, but the platform around it. That obsession would become Nvidia's greatest weapon.
Jensen Huang, Chris Malachowsky, Curtis Priem. Goal: graphics chips for video games. Starting capital: $40,000.
Nvidia coins the term "GPU" (Graphics Processing Unit) with the launch of the GeForce 256. The company goes public. Revenue: $158 million.
Nvidia releases CUDA โ a free programming platform that lets developers write code to run on Nvidia GPUs. At the time, seen as a side project. In hindsight: the most consequential software bet in AI history.
A research team uses Nvidia GPUs to train a neural network called AlexNet. It obliterates the competition in a global image recognition contest. The AI world realizes GPUs are the key to deep learning. Nvidia is already there, with CUDA ready.
Nvidia launches the H100 GPU. ChatGPT explodes. Every AI lab on Earth desperately queues for H100s. Single chips sell for $40,000 on secondary markets. Waitlists stretch 6โ12 months.
Nvidia briefly surpasses Apple and Microsoft to become the most valuable company in the world. Revenue from AI chips: $90 billion. Two years prior it was $15 billion. No company in history grew that fast at that scale.
What Is an AI Chip โ And Why GPUs Became the Engine of Intelligence
To understand why Nvidia rules the AI world, you first need to understand what an AI chip actually does โ and why it's so different from the chip inside your laptop.
A CPU (the chip in your laptop) is like a single brilliant surgeon โ it handles one or a few complex tasks at a time, very fast and very precisely. A GPU is like an army of thousands of simpler workers โ each one isn't as smart, but they all work at the same time in parallel. Training an AI requires doing millions of simple maths operations simultaneously. The GPU's army wins over the surgeon every time for this kind of work.
This is why video game graphics chips turned out to be perfect for AI. Rendering a 3D game frame also requires doing millions of parallel calculations at once โ the same kind of parallel maths that AI needs. Nvidia's GPUs were already built for this. They just needed the right software to unlock it โ and that's what CUDA did.
Imagine Nvidia's GPU is a powerful new type of engine. CUDA is the universal instruction manual that tells every developer in the world how to use that engine. Nvidia released it for free in 2006 โ and for 18 years, every AI researcher, every AI framework, every university AI course has been built on top of it. It's now so deeply embedded in how the whole field works that switching away from Nvidia would mean rewriting everything from scratch.
This is what economists call a "moat." Nvidia's competitors can build a fast chip. Almost nobody can build 18 years of developer trust, tooling, education, and ecosystem. PyTorch, TensorFlow, JAX โ every major AI framework runs on CUDA. That's why Google and Meta, who make their own AI chips, still buy Nvidia chips too.
In 2012, a team of researchers at the University of Toronto used two Nvidia GTX 580 gaming GPUs to train a neural network called AlexNet. It entered the ImageNet competition โ a global contest to identify objects in photos โ and won by a margin so large that the runner-up reportedly thought there was an error. The AI world had its proof of concept. GPUs were the engine. Nvidia was already there. The company's fate was sealed from that day.
Why Nvidia Is Almost Impossible to Beat โ The Full Stack Advantage
Nvidia's dominance isn't just about having the fastest chip. It's about having built something that the entire world's AI infrastructure runs on โ and making it almost impossible to leave.
The Full Stack โ Why Nobody Can Just Copy Nvidia
Nvidia doesn't sell you a chip. It sells you an entire ecosystem. The GPU hardware itself is just the start. On top of that: CUDA โ the programming platform. cuDNN โ a library of AI-optimised math operations. NVLink โ technology that connects thousands of GPUs to act as one giant brain. DGX Servers โ pre-built AI supercomputer boxes ready to deploy out of the box.
A competitor can build a fast chip. Almost nobody in the world can build all five layers of this stack at once โ and have 18 years of developer familiarity on top of it. That's the real fortress.
The network effect compounds this further. Every AI researcher learns on CUDA. Every cloud provider โ AWS, Azure, Google Cloud โ optimises their infrastructure for Nvidia. Every AI startup builds their product assuming Nvidia hardware underneath. Even if a technically superior chip existed tomorrow, the switching cost for the entire industry would be so staggering that Nvidia would still dominate for years.
This is why Jensen Huang โ who still wears his signature black leather jacket to every event โ has been called the most important person in technology. Not because of one great product. Because of thirty years of building an ecosystem that the world cannot function without.
"Nvidia won the same way Windows dominated PCs in the 90s โ not by being the only option, but by becoming the default that everyone built on top of."
Who Is Trying to Challenge Nvidia?
Every major tech company is desperately trying to reduce its dependence on Nvidia. Here's an honest look at where each stands:
The Electricity Crisis โ AI Is Hungry and the Bill Is Terrifying
Here is a number that should make you stop and think: a single Nvidia H100 AI chip uses 700 watts of power. That's roughly the same as running a microwave oven continuously โ just for one chip. A typical AI data centre has tens of thousands of them running simultaneously, 24 hours a day, 7 days a week.
Now you understand why the AI industry has an electricity problem.
A chip is made of billions of tiny switches called transistors. An Nvidia H100 has about 80 billion of them packed into a piece of silicon the size of your palm. Each transistor does one job: flip between ON and OFF โ the 1s and 0s of computing. Every time a transistor flips, it uses a tiny burst of electricity. And some of that energy escapes as heat โ the same reason your phone gets warm when you use it hard. With 80 billion transistors flipping millions of times per second โ the heat and power consumption become enormous.
AI training is worse than regular computing because it runs thousands of calculations simultaneously using thousands of chip cores โ all firing at once, all generating heat, all the time. A regular laptop chip does a few billion operations per second at 15โ45 watts. An H100 does thousands of trillions of operations per second โ at 700 watts. That's the scale difference.
Training GPT-4 is estimated to have consumed as much electricity as 1,000 US homes use in an entire year. Microsoft, Google, and Amazon have all had to quietly reverse their carbon neutrality commitments because AI data centre demand is growing faster than renewable energy can supply.
A single data centre with 50,000 H100 chips running at full power is equivalent to the electrical load of a small city. And it doesn't just stop there โ for every watt a chip consumes, you need roughly another 0.3โ0.5 watts just to cool it down. Giant fans. Chilled water systems. Liquid cooling pumped directly over the chips. All of it costs money, water, and more electricity.
This is the fundamental tension at the heart of the AI boom: the more powerful AI becomes, the more electricity it needs โ and the planet is already struggling to supply it.
The industry is fighting back on multiple fronts. Nvidia's next-gen Blackwell and Vera Rubin chips are designed to deliver dramatically more AI work per watt. Techniques like quantization (using smaller numbers in calculations) and mixture of experts (only activating parts of a model at a time) reduce compute needs. And liquid cooling โ pumping cold water directly over chips โ cuts cooling energy by 30โ40%. But the models keep getting bigger faster than efficiency gains can keep up. The problem is not solved. It is being managed.
Running to the Arctic โ The World's Biggest Free Refrigerator
If your data centre is consuming so much electricity that it generates enough heat to warm a small town โ what do you do? You go somewhere cold.
The tech industry has figured out that the Arctic is essentially the world's largest free refrigerator. And they are flooding into it with extraordinary speed.
In warm climates, data centres need massive, expensive, energy-hungry chiller systems to keep their chips from overheating. In the Arctic, the ambient air is cold enough โ for most of the year โ to cool the data centre using just outside air and natural water sources. No chillers. No refrigerants. Just cold air doing the work for free.
The savings are not small. Cold climate data centres save approximately 30โ40% of their total energy consumption compared to warm-climate equivalents. At the scale of a 100-megawatt facility, that's tens of millions of dollars every year โ just from free cooling.
Southern Finland alone offers about 8,000 hours of free cooling per year. Go further north, and that number climbs higher. Iceland, sitting on volcanic geothermal energy that is essentially free and limitless, is seen as a near-perfect data centre location.
The numbers tell the story of the rush. As of 2025, at least 32 data centres are operating in the Arctic, spread across Arctic nations. Their combined capacity is around 870 megawatts. The single largest belongs to Iceland's Verne Global โ 140 megawatts, enough to power approximately 140,000 homes.
The biggest names in tech are already there. Google operates facilities in Finland. Meta built one of Europe's largest data centres in northern Sweden specifically for the energy and cooling advantages. Microsoft has been evaluating sites in northern Sweden and Finland for massive AI-focused expansions.
One facility planned in northern Norway near Ballangen is designed to draw 100% of its power from wind and hydroelectricity โ and use cold fjord water for passive cooling. Plans call for eventually scaling to over 1,000 megawatts from a single site. That's the equivalent of a large nuclear power plant โ powered entirely by renewables, cooled entirely for free.
Iceland sits on the Mid-Atlantic Ridge โ a zone of intense volcanic activity that gives the country essentially unlimited geothermal energy. No coal. No gas. Just the earth's own heat, converted into electricity for free. Combined with its Arctic temperatures for free cooling, Iceland has become one of the most sought-after locations for data centre development in the world. The only constraint: it's a small island, and space is limited.
The Price Someone Else Pays โ The People the Tech World Forgot to Mention
The Arctic data centre story has a dark side that almost no tech coverage mentions. The Arctic is not empty. It is home to approximately 4 million people โ including Indigenous communities who have lived there, in balance with the land, for thousands of years.
Their names: Alaska Natives, Inuit, First Nations in Canada, Sรกmi in Scandinavia and Russia, and over 40 other ethnic groups speaking more than 40 languages. They didn't cause the AI electricity crisis. But they are paying for it.
Land Displacement: Research in Sรกpmi โ the traditional homeland of the Sรกmi people in Scandinavia โ found that proposed data centre sites directly overlap with reindeer grazing areas. These are not just economic assets. They are cultural and spiritual pillars of Sรกmi life, maintained for thousands of years. Tech companies often describe this land as "underutilised" โ erasing the people entirely from the narrative.
Climate Amplification: The Arctic is already warming at roughly double the rate of the rest of the world. The warming disrupts ice roads, shortens safe travel seasons, and shrinks the habitats of beluga whales, walruses, and seals โ species that Indigenous communities depend on for food and cultural practice. More data centres โ even renewable ones โ add infrastructure and human activity that accelerates this disruption.
Ancient Threats Reawakening: As permafrost thaws, layers of ground frozen for thousands of years are being exposed. In 2016, a 12-year-old boy in Siberia died after being infected with anthrax โ from a long-frozen reindeer carcass exposed during a summer heatwave. Scientists warn that as humans drill and build in previously frozen Arctic ground, they risk encountering viruses and bacteria frozen for up to a million years.
Colonialism by Another Name: Structural inequalities and systemic neglect are directly impacting communities that have been managing this land sustainably for generations. The benefits of AI โ faster chat responses, better image generators, more powerful search engines โ flow to people thousands of miles away. The costs โ disrupted land, accelerated warming, displaced culture โ stay local.
The Indigenous communities of the Arctic did not create the AI electricity problem. They did not demand ChatGPT or image generators or faster search results. But they are watching their ancestral land, their traditional routes, their food sources, and their way of life disappear โ because the rest of the world cannot figure out how to power its AI without running to the coldest place left on earth. That is a story that deserves far more attention than it gets.
Elon Musk's Wildest Bet โ Take the Data Centers to Space
If the problem is that data centres need enormous amounts of electricity and cooling โ and the Arctic is running out of space โ what if you removed the problem entirely? What if you put the data centres somewhere with unlimited free solar power and zero cooling costs?
What if you put them in space?
This sounds like science fiction. It is not. On January 30, 2026, SpaceX filed with the US Federal Communications Commission for approval to launch up to one million satellites to form a megaconstellation of orbital data centres. Shortly after, SpaceX formally merged with Musk's AI company xAI โ drawing his space and AI ventures together into one infrastructure project.
The logic is not crazy. In orbit, solar panels are dramatically more efficient โ no clouds, no weather, no atmosphere absorbing the sunlight. In the right orbit, solar panels can collect energy almost continuously. And there is no air in space โ so no need for complex, expensive, energy-hungry cooling systems. The heat can simply radiate away into the void.
SpaceX already has the beginnings of this infrastructure. Its 8,000 Starlink satellites collectively contain nearly half a million computers powered by solar arrays totalling about 100 megawatts โ roughly the computing count and power used for a large earthbound data centre, spread across orbit, connected by free-space optical laser links.
Musk has said rocket launch costs have fallen so dramatically that within three years, space will be the cheapest way to generate AI compute power. He has set 2028 as the tipping point year.
"You can mark my words: in 36 months โ probably closer to 30 months โ the most economically compelling place to put AI will be space."
โ Elon Musk, 2025Musk went further. He predicted: "Five years from now, we will launch and be operating every year more AI in space than the cumulative total on Earth."
The Problems โ And They Are Very Real
The vision is spectacular. The obstacles are formidable. Here are the ones that matter most:
In May 2025, China launched the first dozen satellites of its planned 2,800-satellite "Three-Body Computing Constellation" โ a literal supercomputer in space, connected by laser communications. Sam Altman has mused that space data centres might be the long-term solution. Jeff Bezos has predicted "giant gigawatt data centres in space" within 20 years. Former Google CEO Eric Schmidt acquired a rocket company as a bet on orbital computing. The race for the sky is real โ and it has geopolitical dimensions beyond just electricity costs.
The Pattern Nobody Wants to Say Out Loud โ Every Solution Has a Hidden Cost
Step back from all of this and you start to see a pattern.
| Location | Advantage | Hidden Cost | |
|---|---|---|---|
| ๐ก๏ธ | Regular Data Centres | Close to users, easy to maintain | Burns massive electricity, water, heats cities, breaks climate pledges |
| ๐ง | Arctic Data Centres | Free cooling, clean energy, 30โ40% savings | Destroys Indigenous land, accelerates Arctic warming, exposes ancient pathogens |
| ๐ | Space Data Centres | Free solar power, no cooling needed, unlimited space | $35B+ extra cost, space junk risk, ozone damage, decades away |
The Question Worth Asking
We started this story with a 30-year-old engineer sketching on a diner napkin. That sketch became the chip that runs every AI on earth. That chip now consumes so much electricity that the world is moving its computers to the Arctic โ displacing Indigenous communities who have lived there for thousands of years. And the next proposed solution is to put a million satellites in orbit and risk making space unusable for generations.
Each step in this chain was rational. Each step had brilliant people behind it. And yet the cumulative result is a technology whose power consumption is actively reversing humanity's climate commitments โ with the costs falling most heavily on people who had nothing to do with creating any of it.
That is the story of the chip that runs the world. And it deserves to be told whole โ not just the triumph, but the price.