Wong Edan's

AI’s Gluttony: Silicon Shortage Deja Vu, But Worse

February 08, 2026 • By Azzar Budiyanto

The AI Tsunami: Drowning in Demand, Starving for Silicon

Alright, listen up, you tech-heads, digital dreamers, and anyone still clinging to the naive hope that our interconnected world won’t eventually eat itself. Your friendly neighborhood tech prophet, Wong Edan, is here to deliver some sobering news with a side of “I told you so.” We’ve barely dusted ourselves off from the last global chip shortage – a nasty little pandemic-induced hangover that messed with everything from your PlayStation 5 to your new car – and guess what? We’re barreling headfirst into another one. And this time, the villain isn’t some invisible virus shutting down factories. No, this time, the culprit is the shiny, seductive, insatiably hungry beast we call Artificial Intelligence. Gila!

The murmurs have been growing louder, analysts are starting to sound the alarm, and frankly, if you’ve been paying attention, you’d know it’s not just a murmur; it’s a full-blown roar. CNBC warned us back in September 2024 that “A surge in demand for AI semiconductors and AI-enabled smartphones and laptops could lead to the next global chip shortage.” Bain & Company followed up, urging us to “Prepare for the Coming AI Chip Shortage.” Prepare? My friends, the preparation phase is over. We’re in the thick of it, and it’s getting ugly.

This isn’t just a supply chain hiccup. This is a fundamental clash between unprecedented technological ambition and the cold, hard realities of advanced manufacturing. AI isn’t just using chips; it’s devouring them, and it’s spitting out consequences that will impact everything from the price of your next phone to the pace of innovation in every industry imaginable. So, grab a strong coffee, maybe a stiff drink, and let’s dive into the silicon storm that AI has unleashed.

The Ghost of Shortages Past, The Shadow of Shortages Future

Remember 2020-2022? The great chip shortage where everyone suddenly became an expert on “fabs” and “node sizes”? That was a mess. Factories idled, car production stalled, graphic card prices went stratospheric, and your washing machine suddenly had a longer lead time than a custom yacht. We blamed COVID-19, geopolitical tensions, and a sudden, unexpected spike in demand for home electronics. And we learned… well, hopefully, we learned something about fragile global supply chains.

But the AI-driven shortage is different. It’s not a sudden, exogenous shock. It’s a systemic, self-inflicted wound born from our collective obsession with intelligent machines. It’s an economic phenomenon driven by the insatiable appetite of AI models that need more processing power and memory than we ever thought possible. As Bain & Company rightly pointed out, “While businesses couldn’t predict the pandemic, they can guard against the next big threat to semiconductor supply chains.” The irony, of course, is that while we could see this coming, the sheer scale of AI demand has made “guarding against it” feel like trying to stop a tsunami with a sandcastle.

The previous shortage was about general chip capacity – microcontrollers, power management ICs, commodity memory. This new crisis, fueled by the AI frenzy, is far more acute, focused primarily on the absolute bleeding edge of semiconductor technology. We’re talking about incredibly complex, power-hungry, and ludicrously expensive chips that only a handful of companies can design, and even fewer can manufacture. This isn’t just a scarcity; it’s a scarcity of the rarest jewels in the silicon kingdom.

AI’s Insatiable, Gluttonous Appetite: Why So Many Chips?

So, why is AI such a hungry, hungry hippo when it comes to silicon? Let’s break it down, because understanding the “why” is crucial to grasping the magnitude of the problem.

1. Training, Training, and More Training

At the heart of modern AI, especially large language models (LLMs) and generative AI, lies the concept of deep learning. This involves neural networks with billions, even trillions, of parameters. To make these networks “smart,” they need to be trained on colossal datasets – think petabytes of text, images, and video. This training process is computationally brutal. It involves:

  • Massive Parallel Computation: Neural networks are essentially vast arrays of matrix multiplications. GPUs (Graphics Processing Units) excel at this, performing thousands of calculations simultaneously.
  • Iterative Refinement: The model learns through countless iterations, adjusting its parameters based on feedback. Each iteration requires immense processing power.
  • Sheer Scale: Models like GPT-4 or Google’s Gemini aren’t trained on a single GPU. They require clusters of thousands of high-end GPUs, running for weeks or months. Imagine the electricity bill, let alone the hardware cost!

Each training run consumes an astronomical amount of compute. And with new, larger, more complex models emerging every few months, the demand for these top-tier training chips is escalating exponentially.

2. Inference at Scale: AI Everywhere, All at Once

Once an AI model is trained, it needs to be used. This is called “inference” – applying the learned knowledge to new data to make predictions or generate content. Think ChatGPT answering your queries, DALL-E generating an image, or AI-powered search engines. While inference can sometimes run on less powerful hardware than training, when you’re talking about millions or billions of users interacting with AI services simultaneously, the aggregated demand is still immense.

  • Cloud AI: Most sophisticated AI inference happens in vast data centers, often operated by hyperscalers like Amazon, Microsoft, Google, and Meta. These data centers are now being retrofitted and built specifically with AI workloads in mind, meaning they need racks upon racks of AI accelerators.
  • Edge AI: The push for “AI-enabled” devices – smartphones, laptops, smart home devices, self-driving cars – means bringing AI processing closer to the user. This requires specialized Neural Processing Units (NPUs) or other dedicated AI accelerators built directly into consumer hardware. As CNBC noted, “AI-enabled smartphones and laptops” are a significant demand driver.

3. The Memory Monster: HBM and DDR5

Here’s where it gets particularly nasty. It’s not just the processing units; it’s the memory that feeds them. AI models, especially during training, require incredibly fast access to vast amounts of data. This isn’t your grandma’s DDR4 RAM; we’re talking about next-generation memory technologies. As Reuters highlighted in December 2025, “An acute global shortage of memory chips is forcing artificial intelligence and consumer-electronics companies to fight for dwindling supply.”

  • High Bandwidth Memory (HBM): This is the Holy Grail for AI. HBM stacks multiple DRAM dies vertically, connected by through-silicon vias (TSVs) to achieve unprecedented data transfer rates. An Nvidia H100 GPU, for instance, comes packed with multiple HBM3 stacks, providing terabytes per second of memory bandwidth. Without HBM, even the fastest AI processor would be starved for data, severely limiting its performance. The problem? HBM is incredibly complex and expensive to manufacture, requiring specialized packaging techniques and very high yields. Only a few players like SK Hynix, Samsung, and Micron are in this game.
  • DDR5 DRAM: While HBM caters to the high-end, general-purpose DDR5 DRAM is still crucial for systems hosting AI chips. Its higher speeds and capacities compared to DDR4 are a necessity.
  • NAND Flash: AI also churns through vast amounts of data storage. High-performance NAND flash is needed for storing datasets, model weights, and intermediate results. The WSJ warned in January 2026 that “AI companies’ need for a type of once-affordable microchip threatens to drive up prices of all electronics—and limit data-center ambitions.” This refers to both advanced logic and memory chips, particularly the latter.

Samsung, a titan in memory production, foresaw an “acute chip shortage persisting, driven by the AI boom” as early as January 2026, noting that “strong memory demand benefitting its mainstay” – meaning, they’re selling all they can make, and it’s not enough. The AI frenzy is indeed driving a memory chip supply crisis, making prices for all devices potentially rise, as Reuters and other analysts have noted.

Nvidia: The Green Giant and the Golden Goose (or the Bottleneck?)

If AI is the hungry beast, then Nvidia is the primary feeder. Their GPUs, particularly the A100, H100, and soon-to-be B200 “Blackwell” series, are the undisputed champions of AI compute. Nvidia doesn’t just sell hardware; they sell an entire ecosystem built around their CUDA software platform, which makes developing and deploying AI models incredibly efficient on their chips. This has created an almost insurmountable moat around their market dominance. Other companies, like AMD with their MI series and various startups building ASICs (Application-Specific Integrated Circuits), are trying to catch up, but it’s a monumental task.

The problem is, Nvidia can’t print money fast enough to ramp up supply to meet demand. Or rather, they can print money, but they can’t print silicon wafers. Their chips are manufactured by TSMC (Taiwan Semiconductor Manufacturing Company), the world leader in advanced chip fabrication. And TSMC’s cutting-edge nodes (5nm, 4nm, 3nm) are in incredibly high demand from all leading-edge chip designers.

The situation is so tight that even in China, where geopolitical tensions complicate things, there are significant shortages. H3C, one of China’s largest server makers, warned in March 2025 of “potential shortages of Nvidia’s H20 chip, the most advanced AI processor legally available domestically.” The H20 is a toned-down version of Nvidia’s high-end chips, specifically designed to comply with US export restrictions. Yet, even this restricted chip is becoming a bottleneck. This underscores not only the sheer global demand but also how geopolitical maneuvering fragments supply and exacerbates scarcity.

"Nvidia's grip on the AI accelerator market is a double-edged sword. On one hand, it drives innovation and provides the tools for incredible advancements. On the other, it creates a single point of failure in the global AI supply chain. When Nvidia sneezes, the AI world catches a cold – and right now, Nvidia's got a chronic cough of demand."

The HBM Headache: High Bandwidth, Higher Barriers

Let’s talk more about HBM, because it’s a specific, critical choke point that illustrates the problem perfectly. As discussed, HBM is essential for feeding AI GPUs data fast enough. Imagine an AI chip as a super-fast chef, and HBM as the ingredients. If the chef has to wait for ingredients, even if they’re the best chef in the world, they can’t cook efficiently. HBM ensures a constant, rapid flow of data.

Manufacturing HBM is ridiculously complex. It involves:

  1. Stacking DRAM Dies: Imagine stacking multiple layers of ultra-thin memory chips on top of each other.
  2. Through-Silicon Vias (TSVs): These are tiny vertical electrical connections that go through the silicon dies, connecting the layers. They are incredibly difficult to make reliably and efficiently.
  3. Advanced Packaging: The entire HBM stack then needs to be integrated onto the same package as the GPU, often using sophisticated techniques like 2.5D or 3D stacking, requiring highly specialized equipment and expertise.

Only a handful of companies (SK Hynix, Samsung, Micron) have mastered this process, and each has limited capacity. The yield rates for HBM can be challenging, meaning not every manufactured chip stack is perfect. This intricate dance of design, manufacturing, and packaging means that even if Nvidia can get all the GPU dies it wants from TSMC, if there isn’t enough HBM to go with it, those GPUs are essentially useless for AI workloads.

This bottleneck is holding up entire data center expansions and slowing down AI development across the globe. It’s a testament to how specialized and interconnected modern semiconductor manufacturing has become.

Beyond the Server Rack: The Consumer’s Coming Pain

While the headlines often focus on data centers and high-end AI training, the ripple effects of this shortage will inevitably reach your pocket. As Marketplace reported in January 2026, “AI investment continues to drive chip shortage,” impacting not just cutting-edge tech but also “manufacturers of consumer electronics are warning of price increases.”

  • Smartphones and Laptops: The push for “AI-enabled” devices means these consumer gadgets will increasingly demand dedicated AI chips (NPUs) and faster, more efficient memory (like LPDDR5/5X, which competes for fab space with server-grade memory). If the high-end AI chips are sucking up all the manufacturing capacity, it leaves less for the components that go into your everyday devices.
  • Automotive: Remember Honda postponing the reopening of a plant in China? While not directly stated as AI, the automotive industry relies heavily on a multitude of chips, from simple microcontrollers to advanced processors for infotainment and ADAS (Advanced Driver-Assistance Systems), which increasingly incorporate AI. When the global capacity is strained, all chip demand feels the pinch.
  • Price Increases: This is the most direct impact on consumers. When supply shrinks and demand surges, prices go up. “The Global Memory-Chip Shortage Will Cost Us All,” warned the WSJ in January 2026, stating that AI companies’ need for memory chips “threatens to drive up prices of all electronics.” Your next phone, laptop, or even a smart appliance could carry a higher price tag simply because the underlying components are harder and more expensive to source. “Memory loss: As AI gobbles up chips, prices for devices may rise,” echoed Reuters in December 2025. It’s not just a prediction; it’s a foregone conclusion.

This isn’t just about AI itself becoming more expensive; it’s about the entire electronics ecosystem being starved of the silicon it needs. The big fish (AI companies) are eating all the little fish (general consumer electronics components), and everyone else is left hungry.

Geopolitics and the Silicon Iron Curtain

As if global supply and demand weren’t complex enough, throw in a generous helping of geopolitical rivalry, and you’ve got a recipe for pure chaos. The US-China tech war has profoundly impacted the AI chip landscape, exacerbating the global shortage and creating parallel, less efficient ecosystems.

  • Export Bans: In October 2022, the United States banned the export to China of “any AI chips equal to or more capable than the Nvidia A100 chip.” This wasn’t just a slap on the wrist; it was a strategic move to hamstring China’s ability to develop cutting-edge AI, which is seen as critical for future economic and military power. This ban was later expanded to include more chips and tighter restrictions.
  • China’s AI Chip Deficit: The ban has forced Chinese tech giants to either hoard existing chips or seek domestic alternatives. While companies like Huawei are pouring billions into developing their own AI chips (like the Ascend series), they’re still playing catch-up. As China’s AI Chip Deficit highlighted in December 2025, “Huawei Can’t Catch Nvidia and U.S. technologies.” The leading-edge fabs and intellectual property needed to produce chips truly competitive with Nvidia’s H100 are simply not available domestically due to technological gaps and the inability to access advanced lithography equipment from ASML.
  • Fragmented Supply: The result is a fragmented global supply chain. Nvidia has had to create specific, less powerful chips (like the H20 for China) to comply with regulations, tying up manufacturing capacity that could otherwise go to higher-performing chips for the rest of the world. Chinese companies are forced to settle for less, or invest heavily in domestic production that is years behind. This creates inefficiencies and prevents a truly global optimization of chip allocation.
  • The Scramble for Self-Sufficiency: The geopolitical tensions have spurred nations and blocs (US, EU, China, Japan) to invest massively in domestic chip manufacturing capabilities. While this is a long-term strategy for resilience, it adds immense cost and duplicates efforts, potentially leading to regional oversupply in some areas while global shortages persist for the most advanced components. Building a cutting-edge fab costs tens of billions of dollars and takes years. We’re talking about a multi-decade race here.

This isn’t just about silicon; it’s about sovereignty, economic power, and military advantage. And in this high-stakes game, the global chip shortage becomes not just an economic challenge but a geostrategic battlefield.

The Foundry Follies: Building the Future, Brick by Billion

At the very foundation of this entire crisis are the foundries – the companies that actually manufacture these minuscule marvels. TSMC, Samsung Foundry, and to a lesser extent, Intel Foundry, are the titans here. They operate multi-billion dollar fabrication plants (fabs) that are among the most complex engineering feats on the planet.

  • Cost and Time: Building a new cutting-edge fab isn’t like putting up a new factory. We’re talking 10-20 billion dollars and 3-5 years from groundbreaking to mass production. And that’s just for one fab! To significantly increase global capacity for advanced chips would require dozens of these, costing hundreds of billions and taking well over a decade.
  • Advanced Process Nodes: AI chips thrive on the latest process nodes (e.g., TSMC’s 3nm, 2nm). These nodes pack more transistors into a smaller area, making chips faster and more power-efficient. But shrinking transistors further requires incredibly sophisticated and expensive equipment, notably EUV (Extreme Ultraviolet) lithography machines from ASML, which cost hundreds of millions each and are themselves a bottleneck.
  • Yield Rates: Manufacturing at these advanced nodes is incredibly difficult. Even a microscopic speck of dust can ruin a wafer full of chips. Achieving high yield rates (the percentage of functional chips on a wafer) is a constant challenge, and low yields reduce the effective supply even further.
  • The Ecosystem: It’s not just the fab itself. It’s the entire ecosystem: chemical suppliers, equipment manufacturers, specialized engineers, water and power infrastructure. It’s a massive undertaking that cannot be spun up overnight, regardless of how much money is thrown at it.

The foundries are running at full capacity, but AI demand is growing faster than new fabs can come online. This is the fundamental, painful truth. We’re in a vicious cycle where the demand for the most advanced silicon is simply outstripping the world’s ability to produce it.

When Does the Pain Stop? Crystal Ball Gazing (Wong Edan Style)

Ah, the million-dollar question: when will this silicon apocalypse end? My crystal ball is a bit hazy from all the AI smoke, but here’s the Wong Edan take:

It won’t be anytime soon. Samsung, a company that lives and breathes chips, forecasts an “acute chip shortage persisting” throughout 2026 and potentially beyond. Bain & Company’s call to “guard against the next big threat” implies a sustained period of vulnerability.

Why the pessimism? Because the demand side isn’t letting up. AI development is accelerating, not slowing down. Every major tech company, every startup, every research institution is scrambling for more compute. And the supply side? Fabs take years to build. New HBM manufacturing lines take time to qualify and ramp up. The sheer physics and economics of advanced semiconductor manufacturing dictate a slow, arduous path to increased supply.

We might see temporary reprieves or shifts in bottlenecks. Maybe HBM supply improves, but then advanced packaging becomes the constraint. Or perhaps a new generation of AI chips requires a different memory type, shifting the shortage. But the underlying tension between boundless AI ambition and finite physical resources will persist for several years, likely well into the late 2020s.

"Expecting a quick fix for the AI chip shortage is like expecting your toddler to stop asking for snacks after one bite. It’s a marathon, not a sprint, and we're just hitting mile one."

Navigating the Silicon Storm: What Now, Edan?

So, what’s a tech enthusiast, a CEO, or even just a bewildered consumer to do in this silicon storm?

For Industry Leaders & AI Innovators:

  • Diversify, Diversify, Diversify: Relying on a single supplier or a single type of chip (ahem, Nvidia) is a recipe for disaster. Explore alternatives like AMD’s MI series, Intel’s Gaudi accelerators, and custom ASICs where feasible. This won’t eliminate the shortage but can mitigate risk.
  • Optimize and Innovate on Software: Squeeze every last drop of performance from existing hardware. More efficient algorithms, better data compression, and smarter model architectures can reduce the raw compute requirement. Software innovation can sometimes be a stop-gap for hardware scarcity.
  • Vertical Integration & Strategic Partnerships: Companies like Microsoft, Google, and Amazon are already designing their own custom AI chips (like Microsoft’s Maia and Cobalt, Google’s TPUs). This is a long, expensive road but offers greater control over supply. Smaller players need strong, long-term partnerships with foundries and memory manufacturers.
  • Think Beyond Traditional Architectures: Explore neuromorphic computing, quantum computing, or other alternative computing paradigms that might offer a different path to AI without relying solely on traditional silicon. This is speculative, but necessary long-term thinking.
  • Rethink Data Center Design: Focus on energy efficiency, innovative cooling, and modular designs that can adapt to varying chip availabilities.

For Consumers & Tech Enthusiasts:

  • Expect Higher Prices: Brace yourselves. Your next smartphone, laptop, or even smart TV might cost a little (or a lot) more. The trickle-down effect from high-end memory and logic chip scarcity is real.
  • Patience is a Virtue: The latest and greatest “AI-enabled” features might take longer to become mainstream or might command a premium. Product cycles could extend slightly.
  • Appreciate What You Have: That existing hardware still has life in it. Software updates can often bring new “AI features” to older devices without needing a hardware upgrade.
  • Be Wary of Hype: Not every device needs a dedicated NPU. Understand what “AI-enabled” truly means for your specific use case. Sometimes, simpler is better, and less demanding of scarce resources.

Conclusion: The Silicon Squeeze is Here to Stay (For Now)

The bottom line, my friends, is that the AI revolution, while undeniably transformative, comes with a hefty price tag – and that price is being paid in silicon. We are facing a complex, multi-faceted chip shortage driven by an insatiable demand for processing power and high-bandwidth memory, exacerbated by geopolitical tensions and the sheer, unyielding realities of advanced manufacturing.

This isn’t a problem that will magically disappear. It’s a fundamental reordering of the global technology landscape. Companies will adapt, innovate, and find new ways to squeeze performance out of every available wafer. But the days of cheap, abundant, bleeding-edge silicon are, for now, a distant memory. The “Wong Edan” prognosis? Prepare for a bumpy ride, folks. The AI beast is hungry, and it’s eating the world’s chips, one precious transistor at a time. It’s a mess, but by the gods, it’s a fascinating mess to watch unfold. Let’s see who survives the silicon squeeze.