Optimizing In-Memory Computing for Energy Efficiency and Flexibility

The current research landscape in the field of in-memory computing (IMC) and compute-in-memory (CIM) accelerators is witnessing significant advancements aimed at enhancing energy efficiency, flexibility, and performance. Innovations in analog IMC architectures are focusing on integrating shift-add capabilities directly within memory arrays, leveraging emerging non-volatile memory technologies like FeFETs. These advancements not only streamline the MAC operations but also optimize energy and area efficiency by eliminating the need for complex shift-add circuitry. On the other hand, digital CIM accelerators for spiking neural networks (SNNs) are evolving to support flexible operand resolutions and layer-wise weight/output stationarity, enabling significant energy savings and improved inference latency in edge applications. The integration of online alignment and addition algorithms in floating-point adders is also contributing to reduced area and power consumption in multi-term floating-point operations, enhancing the overall efficiency of data aggregation processes.

Noteworthy papers include one that introduces a novel FeFET-based analog IMC design with inherent shift-add capability, significantly improving energy efficiency. Another notable contribution is a digital CIM accelerator for SNNs that offers flexible operand resolution and layer-wise stationarity, achieving substantial energy savings and high classification accuracy.

Sources

Energy Efficient Dual Designs of FeFET-Based Analog In-Memory Computing with Inherent Shift-Add Capability

Online Alignment and Addition in Multi-Term Floating-Point Adders

An Event-Based Digital Compute-In-Memory Accelerator with Flexible Operand Resolution and Layer-Wise Weight/Output Stationarity

Built with on top of