Exploring Extreme Quantization in Spiking Language Models

Malyaban Bal, Yi Jiang, Abhronil Sengupta

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Despite the growing prevalence of large language model (LLM) architectures, a crucial concern persists regarding their energy and power consumption, which still lags far behind the remarkable energy efficiency of the human brain. Recent strides in spiking language models (LM) and transformer architectures aim to address this concern by harnessing the spiking activity of biological neurons to enhance energy/power efficiency. Doubling down on the principles of model quantization and energy efficiency, this paper proposes the development of a novel binary/ternary (1/1.58-bit) spiking LM architecture. Achieving scalability comparable to a deep spiking LM architecture is facilitated by an efficient knowledge distillation technique, wherein knowledge from a non-spiking full-precision 'teacher' model is transferred to an extremely weight quantized spiking 'student' LM. Our proposed model represents a significant advancement as the first-of-its-kind 1/1.58-bit spiking LM, and its performance is rigorously evaluated on multiple text classification tasks of the GLUE benchmark.

Original languageEnglish (US)
Title of host publicationProceedings - 2024 International Conference on Neuromorphic Systems, ICONS 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages272-276
Number of pages5
ISBN (Electronic)9798350368659
DOIs
StatePublished - 2024
Event2024 International Conference on Neuromorphic Systems, ICONS 2024 - Arlington, United States
Duration: Jul 30 2024Aug 2 2024

Publication series

NameProceedings - 2024 International Conference on Neuromorphic Systems, ICONS 2024

Conference

Conference2024 International Conference on Neuromorphic Systems, ICONS 2024
Country/TerritoryUnited States
CityArlington
Period7/30/248/2/24

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence
  • Computer Science Applications
  • Computer Vision and Pattern Recognition
  • Signal Processing
  • Modeling and Simulation

Fingerprint

Dive into the research topics of 'Exploring Extreme Quantization in Spiking Language Models'. Together they form a unique fingerprint.

Cite this