Pilot Talk 1: In-Memory Computing based Machine Learning Accelerators: Opportunities and Challenges

Pilot Talk 1: In-Memory Computing based Machine Learning Accelerators: Opportunities and Challenges

3.566 Lượt nghe
Pilot Talk 1: In-Memory Computing based Machine Learning Accelerators: Opportunities and Challenges
NSF-PIM Pilot Talk 1: Dr. Kaushik Roy Professor in Purdue University 00:00:00 NSF-PIM Introduction 00:03:53 Pilot Talk1 00:11:40 Pilot Talk1: Background In-Memory Computing 00:19:43 Pilot Talk1: Non-volatile Memory Crossbars 00:30:44 Pilot Talk1: Challenges: NVM devices 00:36:33 Pilot Talk1: Challenges: Architecture 00:38:28 Pilot Talk1: Potential Solutions 01:07:06 Pilot Talk1: Q&A Abstract Machine Learning applications, especially Deep Neural Networks (DNNs) have seen ubiquitous use in computer vision, speech recognition, and robotics. However, the growing complexity of DNN models have necessitated efficient hardware implementations. The key compute primitives of DNNs are matrix vector multiplications which leads to significant data movement between memory and processing units in today’s von Neumann systems. A promising alternative would be co-locating memory and processing elements, which can be further extended to performing computations inside the memory itself. We believe in-memory computing is a propitious candidate for future DNN accelerators since it mitigates the memory wall bottleneck. In this talk, I will discuss various in-memory computing primitives in both CMOS and emerging non-volatile memory (NVM) technologies. Subsequently, I will describe how such primitives can be incorporated in stand-alone machine learning accelerator architectures. Finally, I will focus on the challenges associated with designing such in-memory computing accelerators and explore future opportunities.