Memory is Breaking Product Plans
- Memory now a critical constraint
- AI has disproportionate impact
- Late design engagement reduces options
Over the past few quarters, memory has shifted from a routine sourcing step to a critical design and business constraint—often surfacing at the worst possible moment, just as products move from development to production.
Engineering teams are discovering that memory, once viewed as stable and low risk, is now reshaping the bill of materials. Procurement teams are learning that established buying habits no longer match current lead times. Suppliers are asking for earlier visibility and firmer commitments, while common substitutions are increasingly unavailable when problems arise.
The pressure point is clear. The tightest constraints are in dynamic random-access memory (DRAM), particularly server-grade DRAM and high-bandwidth memory (HBM). That pressure is spreading across the broader memory ecosystem. Artificial intelligence (AI) infrastructure is consuming a disproportionate share of global capacity. As suppliers prioritize AI, availability tightens across standard double data rate (DDR) and beyond. Quote windows are shrinking, pricing is moving closer to shipment, and predictability is eroding.
This tightening is structural. HBM consumes significantly more wafer capacity per gigabyte than standard DRAM, so each AI accelerator removes outsized capacity from the mainstream supply. What many teams expected to be a short-term imbalance is becoming a longer-term constraint, making assumptions about “near-term normalization” increasingly risky.
While DRAM is the most visible issue, pressure now extends across NAND, NOR, managed flash (eMMC, UFS), memory modules, embedded boards as well as adjacent technologies that share wafer, packaging and test capacity.
NAND follows a different—but equally challenging—path. Rather than AI pulling capacity away, manufacturers are actively managing supply alignment. As a result, managed flash devices like eMMC, UFS and SSDs remain exposed to allocation risk even when end demand appears soft.
So what?
For OEM teams, the takeaway is immediate: the market has moved from availability on demand to allocation by commitment. Treating memory as a spot market commodity now carries real risk—jeopardizing timelines, margins and even product viability.
Teams navigating this environment successfully are not waiting for conditions to normalize. They are committing earlier, designing with practical flexibility, and aligning demand signals with actual consumption rather than optimistic forecasts.
The early warning signs
In hindsight, the warning signs weren’t dramatic. They showed up quietly in operations.
Inventory availability tightened first. Pricing behavior followed, becoming less predictable—often turning into “price of the day,” with final pricing set at shipment rather than order placement.
Supplier expectations also shifted. Forecasts mattered less unless backed by real commitments. Long-term agreements and non-cancelable, non-returnable terms became more common, especially for customers seeking meaningful volume.
Today, allocation is driven by tangible demand: purchase orders, backlog, and long-term commitments. Customers receiving supply now are often those who ordered six to eight months ago, reflecting current lead time realities.
Overforecasting, however, is risky. It may gain attention in the short term, but it damages credibility when orders don’t come through. In an allocation environment, credibility is currency—and once lost, it’s hard to regain.
This dynamic is especially difficult for small- and mid-size OEMs. Larger customers can absorb more risk and commit at scale. Others must rely on earlier planning and disciplined demand signaling.
For many teams, the shortage only became real when costs spiked. A typical embedded BOM might include $5 each for the processor, flash, RAM and miscellaneous components. When RAM jumps from $5 to $20, the economics of the entire product change. The question quickly shifts from “Can we source this?” to “Does this product still make sense?”
A common pattern often occurs. Procurement initially escalates the issue, trying to resolve it through sourcing. Only later—sometimes weeks afterward, when it becomes clear that pricing and availability can’t be fixed without revisiting design assumptions—does engineering get involved.
The memory squeeze, quantified
DRAM Q1 2026
Contract prices up 90-95% QoQ (Trendforce, February 2026)
PC DRAM Q1 2026
Contract prices up 100%+ QoQ (record) (TrendForce, February 2026)
NAND Q1 2026
Contract prices up 55-60% QoQ; client SSD prices up 40%+ (TrendForce, February 2026)
Lead times
25-45+ weeks (vs. 8-12 weeks historical norm) (SHI, February 2026)
HBM wafer consumption
4x standard DRAM per gigabyte (Tom’s Hardware, December 2025)
Why memory problems become program problems
One persistent misconception is that memory can be swapped easily. In many designs, it cannot.
Memory choice is often dictated by the processor architecture and its memory controller. While switching suppliers within the same DDR generation may be feasible, moving between DDR generations is not. Pinouts, signaling, timing and validation all differ. Memory decisions are frequently architecture decisions—not sourcing decisions.
This reality is driving choices that once seemed counterintuitive. Some OEMs are intentionally selecting older processors because they pair with older, more affordable or more available memory. The logic is simple: it works, it’s sufficient for the near term and the transition can be managed later.
That trade-off explains why redesign enters the conversation so quickly. If memory is constrained or prohibitively expensive, options narrow fast: wait, pay or redesign. And redesign is rarely trivial.
In long lifecycle, regulated markets—medical, military, test and measurement, and industrial systems—change is especially costly. Platforms often remain stable for five to 10 years. In these environments, a memory change can trigger revalidation, requalification or regulatory review, with costs far exceeding the price difference between memory devices.
What suppliers are prioritizing now
Fab expansion takes years, not quarters. With capacity constrained, suppliers are making deliberate choices about where to invest. They are prioritizing products that are both more profitable and aligning with leading-edge technologies, while de-emphasizing or exiting older ones.
In practical terms, when faced with DDR5 versus DDR4, suppliers are choosing DDR5. If they can build this or that, they are building that.
Relief isn’t imminent
Everyone wants to know when the memory market will return to normal. The uncomfortable answer is that “normal” may still be years away. With DRAM, NAND and HBM capacity essentially sold out through 2026, the near-term outlook points to further tightening, not relief. DRAM is expected to remain constrained through mid-2026 before stabilizing gradually, while NAND faces a deeper imbalance that may not resolve until late 2027—pushing memory risk well beyond today’s design and procurement horizons.
The new planning reality
DRAM and NAND are recovering on different timelines. DRAM shows early signs of stabilization later in 2026, while NAND is expected to stay constrained longer—especially for products with high flash content. Memory risk will vary, requiring product-specific strategies instead of broad assumptions.
Supplier guidance is increasingly direct. Supply hasn’t disappeared, but lead times and pricing have changed materially.
For LPDDR4 and DDR4, lead times are extending and becoming less predictable. Allocation is often quarterly, favoring customers with purchase history, early orders and pricing flexibility. Lead times now exceed 25 weeks and can approach 40 weeks. Pricing has doubled in recent months and may continue rising into 2027.
Customers receiving supply today are typically those who placed orders months ago. New entrants are finding little to no short-term inventory.
Pricing models add further complexity. In some cases, buying inventory at net cost can stabilize pricing. In others, pricing is set at the time of shipment, not at the time of the order. Waiting becomes a gamble, but buying without a plan introduces its own risk. Disciplined, early planning is the only sustainable path.
What to do now
When OEM teams ask what to do next, Avnet’s guidance consistently shifts the focus from chasing part numbers to program level actions that reduce risk.
- Start earlier in the design cycle
The greatest leverage exists before the processor and memory subsystem are locked. Late engagement collapses options. - Design with realistic flexibility
Flexibility means intentional, bounded options such as density headroom or multiple suppliers where platforms allow. - Make forecasts defensible
Suppliers prioritize customers with demand signals that match reality. Credibility matters. - Commit earlier
Ordering close to need no longer aligns with memory lead times. Strong outcomes typically involve orders placed months in advance, often with a backlog of nearly a year for critical memory. - Plan by product lifecycle
A three-year product and a 20-year product require different strategies. Long-life products demand more conservative planning. - Use roadmaps to reduce risk
New fab announcements don’t guarantee supply for every product. Aligning designs with technologies that have a clearer long-term runway can reduce risk even if cost pressure remains.
The bottom line
The most reliable lesson is simple: late engagement ruins options. When teams delay, choices narrow to waiting, paying, or redesigning. Engaging earlier—while architecture remains flexible and demand can be turned into credible commitments—significantly improves outcomes.
The memory shortage highlights an increasingly evident reality: supply chain outcomes are increasingly influenced by design choices. Teams that handle memory as a key design component—planning ahead, designing intentionally, and setting realistic commitments—are much better prepared than those waiting for a return to “normal.”