Seminars

View all Seminars  |  Download ICal for this event

Energy-efficient 2.5D Architectures with Processing-in-memory for Machine Learning Applications

Series: CSA Faculty Colloquium

Speaker: Sumit Kumar Mandal, Assistant Professor, Dept. of CSA

Date/Time: Jan 20 16:30:00

Location: CSA Seminar Hall (Room No. 254, First Floor)

Abstract:
Processing-in-memory (PIM) is a promising technique to accelerate deep learning (DL) workloads. Emerging DL workloads (e.g., ResNet with 152 layers) consist of millions of parameters, which increase the area and fabrication cost of monolithic PIM accelerators. The fabrication cost challenge can be addressed by 2.5-D systems integrating multiple PIM chiplets connected through a network-on-package (NoP). However, server-scale scenarios simultaneously execute multiple compute-heavy DL workloads, leading to significant inter-chiplet data volume. State-of-the-art NoP architectures proposed in the literature do not consider the nature of DL workloads. In this talk, we will discuss a novel server-scale 2.5-D manycore architecture that accounts for the traffic characteristics of DL applications. Comprehensive experimental evaluations with different system sizes as well as diverse emerging DL workloads demonstrate that the architecture achieves significant performance and energy consumption improvements with much lower fabrication cost than state-of-the-art NoP topologies.

Speaker Bio:
Sumit Kumar Mandal is currently an assistant professor in Indian Institute of Science, Bangalore. He completed his PhD from University of Wisconsin-Madison. He received best paper award from ACM TODAES in 2020 and ESweek in 2022. His research interest is energy efficient communication architecture for machine learning applications with emerging technologies.

Host Faculty: Prof. Arkaprava Basu