The field of computer architecture is witnessing significant advancements in compute-in-memory and accelerator technologies. Recent developments have led to improved performance, energy efficiency, and scalability in various applications, including artificial intelligence, machine learning, and high-performance computing. Compute-in-memory architectures, such as those using Processing-in-Memory (PIM) and Compute-Express Link (CXL), have shown promising results in reducing data movement and increasing processing efficiency. Additionally, accelerator technologies like GPUs, FPGAs, and specialized ASICs are being designed to optimize specific workloads and improve overall system performance. Noteworthy papers in this area include CIMPool, which proposes a CIM-aware compression and acceleration framework for neural networks, and MVDRAM, which enables GeMV execution in unmodified DRAM for low-bit LLM acceleration. These innovations have the potential to transform the way we design and optimize computing systems, enabling faster, more efficient, and more scalable processing of complex workloads.
Advancements in Compute-in-Memory and Accelerator Technologies
Sources
Improved algorithms for single machine serial-batch scheduling to minimize makespan and maximum cost
A batch production scheduling problem in a reconfigurable hybrid manufacturing-remanufacturing system
MEEK: Re-thinking Heterogeneous Parallel Error Detection Architecture for Real-World OoO Superscalar Processors
FireGuard: A Generalized Microarchitecture for Fine-Grained Security Analysis on OoO Superscalar Cores
HH-PIM: Dynamic Optimization of Power and Performance with Heterogeneous-Hybrid PIM for Edge AI Devices