TY - GEN
T1 - Prediction in the Presence of Response-Dependent Missing Labels
AU - Song, Hyebin
AU - Raskutti, Garvesh
AU - Willett, Rebecca
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/7/11
Y1 - 2021/7/11
N2 - In various settings, limitations of sensing technologies or other sampling mechanisms result in missing labels, where the likelihood of a missing label is an unknown function of the data. For example, satellites used to detect forest fires cannot sense fires below a certain size threshold. In such cases, training datasets consist of positive and pseudo-negative observations (true negatives or undetected positives with small magnitudes). We develop a new methodology and non-convex algorithm which jointly estimates the magnitude and occurrence of events, utilizing prior knowledge of the detection mechanism. We provide conditions under which our model is identifiable. We prove that even though our approach leads to a non-convex objective, any local minimizer has an optimal statistical error (up to a log term) and the projected gradient descent algorithm has geometric convergence rates. We demonstrate on both synthetic data and a California wildfire dataset that our method outperforms existing state-of-the-art approaches.
AB - In various settings, limitations of sensing technologies or other sampling mechanisms result in missing labels, where the likelihood of a missing label is an unknown function of the data. For example, satellites used to detect forest fires cannot sense fires below a certain size threshold. In such cases, training datasets consist of positive and pseudo-negative observations (true negatives or undetected positives with small magnitudes). We develop a new methodology and non-convex algorithm which jointly estimates the magnitude and occurrence of events, utilizing prior knowledge of the detection mechanism. We provide conditions under which our model is identifiable. We prove that even though our approach leads to a non-convex objective, any local minimizer has an optimal statistical error (up to a log term) and the projected gradient descent algorithm has geometric convergence rates. We demonstrate on both synthetic data and a California wildfire dataset that our method outperforms existing state-of-the-art approaches.
UR - https://www.scopus.com/pages/publications/85113484054
UR - https://www.scopus.com/pages/publications/85113484054#tab=citedBy
U2 - 10.1109/SSP49050.2021.9513750
DO - 10.1109/SSP49050.2021.9513750
M3 - Conference contribution
AN - SCOPUS:85113484054
T3 - IEEE Workshop on Statistical Signal Processing Proceedings
SP - 451
EP - 455
BT - 2021 IEEE Statistical Signal Processing Workshop, SSP 2021
PB - IEEE Computer Society
T2 - 21st IEEE Statistical Signal Processing Workshop, SSP 2021
Y2 - 11 July 2021 through 14 July 2021
ER -