TY - JOUR
T1 - Heterogeneous integration for artificial intelligence
T2 - Challenges and opportunities
AU - Mukhopadhyay, S.
AU - Yalamanchili, S.
AU - Swaminathan, M.
AU - Long, Y.
AU - Mudassar, B.
AU - Nair, C. S.
AU - Deprospo, B. H.
AU - Torun, H. M.
AU - Kathaperumal, M.
AU - Smet, V.
AU - Kim, D.
N1 - Funding Information:
This work was supported in part by the National Science Foundation (CNS # 1533767, CCF # 1740197), Semiconductor Research Corporation (#2762.003), Defense Advanced Research Project Agency (# HR0011-17-2-0045), Industry Membership Consortium at the Packaging Research Center, Georgia Tech, and the NSF Center for Advanced Electronics through Machine Learning (CAEML). The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of the NSF, SRC, DARPA, or Georgia Tech.
Publisher Copyright:
© 1957-2012 IBM.
PY - 2019/11/1
Y1 - 2019/11/1
N2 - The recent progress in artificial intelligence (AI) and machine learning (ML) has enabled computing platforms to solve highly complex difficult problems in computer vision, robotics, finance, security, and science. The algorithmic progress in AI/ML have motivated new research in hardware accelerators. The dedicated accelerators promise high energy efficiency compared to software solutions using CPU. However, as AI/ML models become complex, the increasing memory demands and, hence, high energy/time cost of communication between logic and memory possess a major challenge to energy efficiency. We review the potential of heterogeneous integration in addressing the preceding challenge and present different approaches to leverage heterogeneous integration for energy-efficient AI platforms. First, we discuss packaging technologies for efficient chip-to-chip communication. Second, we present near-memory-processing architecture for AI accelerations that leverages 3D die-stacking. Third, processing-in-memory architectures using heterogeneous integration of CMOS and embedded non-volatile memory are presented. Finally, the article presents case studies that integrate preceding concepts to advance AI/ML hardware platform for different application domains.
AB - The recent progress in artificial intelligence (AI) and machine learning (ML) has enabled computing platforms to solve highly complex difficult problems in computer vision, robotics, finance, security, and science. The algorithmic progress in AI/ML have motivated new research in hardware accelerators. The dedicated accelerators promise high energy efficiency compared to software solutions using CPU. However, as AI/ML models become complex, the increasing memory demands and, hence, high energy/time cost of communication between logic and memory possess a major challenge to energy efficiency. We review the potential of heterogeneous integration in addressing the preceding challenge and present different approaches to leverage heterogeneous integration for energy-efficient AI platforms. First, we discuss packaging technologies for efficient chip-to-chip communication. Second, we present near-memory-processing architecture for AI accelerations that leverages 3D die-stacking. Third, processing-in-memory architectures using heterogeneous integration of CMOS and embedded non-volatile memory are presented. Finally, the article presents case studies that integrate preceding concepts to advance AI/ML hardware platform for different application domains.
UR - http://www.scopus.com/inward/record.url?scp=85075108811&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85075108811&partnerID=8YFLogxK
U2 - 10.1147/JRD.2019.2947373
DO - 10.1147/JRD.2019.2947373
M3 - Article
AN - SCOPUS:85075108811
SN - 0018-8646
VL - 63
JO - IBM Journal of Research and Development
JF - IBM Journal of Research and Development
IS - 6
M1 - 8869909
ER -