TY - GEN
T1 - Maximizing Sub-Array Resource Utilization in Digital Processing-in-Memory
T2 - 35th Edition of the Great Lakes Symposium on VLSI 2025, GLSVLSI 2025
AU - Aragonda, Gamana
AU - Najafi, Deniz
AU - Vungarala, Deepak
AU - Tabrizchi, Sepehr
AU - Roohi, Arman
AU - Angizi, Shaahin
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/6/29
Y1 - 2025/6/29
N2 - In digital Processing-in-Memory (PIM) architectures for Artificial Intelligence (AI) acceleration, efficient resource utilization and sub-array level computation reuse remain critical challenges. To address this, mapping strategies such as Bit-Parallel Mapping (BPM) and Bit-Serial Mapping (BSM) have been proposed, each offering distinct latency and parallelism trade-offs. Despite demonstrating promising results in their respective applications, these mapping schemes are hindered by memory inefficiencies. Such wastage reduces the effective memory utilization, limits the number of executable operations, and constrains overall computational throughput. To overcome these limitations, this paper introduces a novel mapping approach, termed U-shaped Mapping (USM), wherein results are organized in a U-shaped pattern within memory arrays. This structure enables efficient data reuse, leading to fast, area-efficient, and high-throughput PIM operations. The proposed technique enables computation over larger input sizes by efficiently utilizing previously unused memory regions. This results in a significant increase in the number of operations that can be executed within a fixed hardware footprint. Experimental evaluation shows that the USM achieves up to 2.33 × reduction in EDP and 2.78 × improvement in throughput compared to the BSM approach. These improvements confirm USM's suitability for optimizing computational efficiency and resource usage in energy- and area-constrained edge devices.
AB - In digital Processing-in-Memory (PIM) architectures for Artificial Intelligence (AI) acceleration, efficient resource utilization and sub-array level computation reuse remain critical challenges. To address this, mapping strategies such as Bit-Parallel Mapping (BPM) and Bit-Serial Mapping (BSM) have been proposed, each offering distinct latency and parallelism trade-offs. Despite demonstrating promising results in their respective applications, these mapping schemes are hindered by memory inefficiencies. Such wastage reduces the effective memory utilization, limits the number of executable operations, and constrains overall computational throughput. To overcome these limitations, this paper introduces a novel mapping approach, termed U-shaped Mapping (USM), wherein results are organized in a U-shaped pattern within memory arrays. This structure enables efficient data reuse, leading to fast, area-efficient, and high-throughput PIM operations. The proposed technique enables computation over larger input sizes by efficiently utilizing previously unused memory regions. This results in a significant increase in the number of operations that can be executed within a fixed hardware footprint. Experimental evaluation shows that the USM achieves up to 2.33 × reduction in EDP and 2.78 × improvement in throughput compared to the BSM approach. These improvements confirm USM's suitability for optimizing computational efficiency and resource usage in energy- and area-constrained edge devices.
UR - https://www.scopus.com/pages/publications/105017554151
UR - https://www.scopus.com/pages/publications/105017554151#tab=citedBy
U2 - 10.1145/3716368.3735277
DO - 10.1145/3716368.3735277
M3 - Conference contribution
AN - SCOPUS:105017554151
T3 - Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI
SP - 817
EP - 822
BT - GLSVLSI 2025 - Proceedings of the Great Lakes Symposium on VLSI 2025
PB - Association for Computing Machinery
Y2 - 30 June 2025 through 2 July 2025
ER -