TY - GEN
T1 - In-memory Realization of In-situ Few-shot Continual Learning with a Dynamically Evolving Explicit Memory
AU - Karunaratne, G.
AU - Hersche, M.
AU - Langeneager, J.
AU - Cherubini, G.
AU - Gallo, M. Le
AU - Egger, U.
AU - Brew, K.
AU - Choi, S.
AU - Ok, I.
AU - Silvestre, C.
AU - Li, N.
AU - Saulnier, N.
AU - Chan, V.
AU - Ahsan, I.
AU - Narayanan, V.
AU - Benini, L.
AU - Sebastian, A.
AU - Rahimi, A.
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Continually learning new classes from few training examples without forgetting previous old classes demands a flexible architecture with an inevitably growing portion of storage, in which new examples and classes can be incrementally stored and efficiently retrieved. One viable architectural solution is to tightly couple a stationary deep neural network to a dynamically evolving explicit memory (EM). As the centerpiece of this architecture, we propose an EM unit that leverages energy-efficient in-memory compute (IMC) cores during the course of continual learning operations. We demonstrate for the first time how the EM unit can physically superpose multiple training examples, expand to accommodate unseen classes, and perform similarity search during inference, using operations on an IMC core based on phase-change memory (PCM). Specifically, the physical superposition of few encoded training examples is realized via in-situ progressive crystallization of PCM devices. The classification accuracy achieved on the IMC core remains within a range of 1.28%-2.5% compared to that of the state-of-The-Art full-precision baseline software model on both the CIFAR-100 and miniImageNet datasets when continually learning 40 novel classes (from only five examples per class) on top of 60 old classes.
AB - Continually learning new classes from few training examples without forgetting previous old classes demands a flexible architecture with an inevitably growing portion of storage, in which new examples and classes can be incrementally stored and efficiently retrieved. One viable architectural solution is to tightly couple a stationary deep neural network to a dynamically evolving explicit memory (EM). As the centerpiece of this architecture, we propose an EM unit that leverages energy-efficient in-memory compute (IMC) cores during the course of continual learning operations. We demonstrate for the first time how the EM unit can physically superpose multiple training examples, expand to accommodate unseen classes, and perform similarity search during inference, using operations on an IMC core based on phase-change memory (PCM). Specifically, the physical superposition of few encoded training examples is realized via in-situ progressive crystallization of PCM devices. The classification accuracy achieved on the IMC core remains within a range of 1.28%-2.5% compared to that of the state-of-The-Art full-precision baseline software model on both the CIFAR-100 and miniImageNet datasets when continually learning 40 novel classes (from only five examples per class) on top of 60 old classes.
UR - http://www.scopus.com/inward/record.url?scp=85141480300&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85141480300&partnerID=8YFLogxK
U2 - 10.1109/ESSCIRC55480.2022.9911329
DO - 10.1109/ESSCIRC55480.2022.9911329
M3 - Conference contribution
AN - SCOPUS:85141480300
T3 - ESSCIRC 2022 - IEEE 48th European Solid State Circuits Conference, Proceedings
SP - 105
EP - 108
BT - ESSCIRC 2022 - IEEE 48th European Solid State Circuits Conference, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 48th IEEE European Solid State Circuits Conference, ESSCIRC 2022
Y2 - 19 September 2022 through 22 September 2022
ER -