TY - GEN
T1 - Trends and opportunities for sram based in-memory and near-memory computation
AU - Srinivasa, Srivatsa
AU - Ramanathan, Akshay Krishna
AU - Sundaram, Jainaveen
AU - Kurian, Dileep
AU - Gopal, Srinivasan
AU - Jain, Nilesh
AU - Srinivasan, Anuradha
AU - Iyer, Ravi
AU - Narayanan, Vijaykrishnan
AU - Karnik, Tanay
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/4/7
Y1 - 2021/4/7
N2 - Changes in application trends along with increasing number of connected devices have led to explosion in the amount of data being generated every single day. Computing systems need to efficiently process these huge amounts of data and generate results, classify objects, stream high quality videos and so on. In-Memory Computing and Near-Memory Computing have been emerged as the popular design choices to address the challenges in executing the above-mentioned tasks. Through In-Memory Computing, SRAM Banks can be repurposed as compute engines while performing Bulk Boolean operations. Near-Memory techniques have shown promise in improving the performance of Machine learning tasks. By carefully understanding the design we describe the opportunities towards amalgamating both these design techniques for obtaining further performance enhancement and achieving lower power budget while executing fundamental Machine Learning primitives. In this work, we take the example of Sparse Matrix Multiplication, and design an I-NMC accelerator which speeds up the index handling by 10x-60x and 10x-70x energy efficiency based on the workload dimensions as compared with non I-NMC solution occupying just 1% of the overall hardware area.
AB - Changes in application trends along with increasing number of connected devices have led to explosion in the amount of data being generated every single day. Computing systems need to efficiently process these huge amounts of data and generate results, classify objects, stream high quality videos and so on. In-Memory Computing and Near-Memory Computing have been emerged as the popular design choices to address the challenges in executing the above-mentioned tasks. Through In-Memory Computing, SRAM Banks can be repurposed as compute engines while performing Bulk Boolean operations. Near-Memory techniques have shown promise in improving the performance of Machine learning tasks. By carefully understanding the design we describe the opportunities towards amalgamating both these design techniques for obtaining further performance enhancement and achieving lower power budget while executing fundamental Machine Learning primitives. In this work, we take the example of Sparse Matrix Multiplication, and design an I-NMC accelerator which speeds up the index handling by 10x-60x and 10x-70x energy efficiency based on the workload dimensions as compared with non I-NMC solution occupying just 1% of the overall hardware area.
UR - http://www.scopus.com/inward/record.url?scp=85106041976&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85106041976&partnerID=8YFLogxK
U2 - 10.1109/ISQED51717.2021.9424263
DO - 10.1109/ISQED51717.2021.9424263
M3 - Conference contribution
AN - SCOPUS:85106041976
T3 - Proceedings - International Symposium on Quality Electronic Design, ISQED
SP - 547
EP - 552
BT - Proceedings of the 22nd International Symposium on Quality Electronic Design, ISQED 2021
PB - IEEE Computer Society
T2 - 22nd International Symposium on Quality Electronic Design, ISQED 2021
Y2 - 7 April 2021 through 9 April 2021
ER -