@inproceedings{d7234b1f2a1645ce9e5e7d3f13168339,
title = "Backdoor Inversion in Neural-Activation Space",
abstract = "There are a variety of defenses against backdoor attacks planted in deep neural network (DNN) classifiers via poisoning of the training set. Backdoor-agnostic methods seek to detect and/or mitigate backdoors irrespective of the incorporation mechanism used by the attacker, while inversion methods explicitly assume one. We describe a new detector that: relies on embedded feature representations (neural-activation space) to estimate (invert) the backdoor and to identify its target class; can operate without access to the training set; and is highly effective for various incorporation mechanisms. Our approach is evaluated - and found favorable - in comparison with an array of published defenses for a variety of attacks.",
author = "Guangmingmei Yang and Xi Li and Hang Wang and George Kesidis and Miller, \{David J.\}",
note = "Publisher Copyright: {\textcopyright} 2025 IEEE.; 35th IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2025 ; Conference date: 31-08-2025 Through 03-09-2025",
year = "2025",
doi = "10.1109/MLSP62443.2025.11204213",
language = "English (US)",
series = "IEEE International Workshop on Machine Learning for Signal Processing, MLSP",
publisher = "IEEE Computer Society",
booktitle = "35th IEEE International Workshop on Machine Learning for Signal Processing",
address = "United States",
}