ChipPub

ChipPub

In-Memory Computing Chips Gain Momentum

In-memory computing slashes edge-AI energy by 62 %—RRAM + ferroelectric hybrids train & infer on-chip

Meng Li's avatar
Meng Li
Oct 26, 2025
∙ Paid

“ChipPub” Publication: 20% Discount Offer Link.


Researchers focus on limiting data movement to reduce power consumption and latency in edge devices.

In popular media, “artificial intelligence” often means large language models running in expensive and power-hungry data centers. However, for many applications, small models running on local hardware are more appropriate.

Autonomous vehicles require real-time responses, avoiding data transmission delays. Medical and industrial applications often rely on sensitive data that cannot be shared with third parties. Although edge AI applications are faster and more secure, their computational resources are extremely limited. They cannot have terabyte-scale memory space or nearly unlimited computing power.

For data centers, these limitations may seem abstract, but they impose strict constraints on edge AI. In an invited paper at the 2025 IEEE International Memory Workshop and its subsequent preprint, Onur Mutlu, a professor of computer science at ETH Zurich, and his colleagues pointed out that in typical mobile workloads, data movement in memory accounts for a shocking 62% of total energy consumption. Memory is the largest consumer of hardware resources, far ahead, but memory latency is often the biggest contributor to execution time.

User's avatar

Continue reading this post for free, courtesy of Meng Li.

Or purchase a paid subscription.
© 2026 Meng Li · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture