TY - GEN
T1 - Constraint-based, transductive learning for distributed ensemble classification
AU - Miller, David J.
AU - Pal, Siddharth
AU - Wang, Yue
PY - 2006/1/1
Y1 - 2006/1/1
N2 - We consider ensemble classification when there is no common labeled data for designing the function which aggregates classifier decisions. In recent work, we dubbed this problem distributed ensemble classification, addressing e.g. when local classifiers are trained on different (e.g. proprietary, legacy) databases or operate on different sensing modalities. Typically, fixed (untrained) rules of classifier combination such as voting methods are used in this case. However, these may perform poorly, especially when the local class priors, used in training, differ from the true (test batch) priors. Alternatively, we proposed a transductive strategy, optimizing the combining rule for an objective function measured on the test batch. We proposed both maximum likelihood (ML) and information-theoretic (IT) objectives and found that IT achieved superior performance. Here, we identify that the fundamental advantage of the IT method is its ability to properly account for statistical redundancy in the ensemble. We also develop an extension of IT that improves its performance. Experiments are conducted on the UC Irvine machine learning repository.
AB - We consider ensemble classification when there is no common labeled data for designing the function which aggregates classifier decisions. In recent work, we dubbed this problem distributed ensemble classification, addressing e.g. when local classifiers are trained on different (e.g. proprietary, legacy) databases or operate on different sensing modalities. Typically, fixed (untrained) rules of classifier combination such as voting methods are used in this case. However, these may perform poorly, especially when the local class priors, used in training, differ from the true (test batch) priors. Alternatively, we proposed a transductive strategy, optimizing the combining rule for an objective function measured on the test batch. We proposed both maximum likelihood (ML) and information-theoretic (IT) objectives and found that IT achieved superior performance. Here, we identify that the fundamental advantage of the IT method is its ability to properly account for statistical redundancy in the ensemble. We also develop an extension of IT that improves its performance. Experiments are conducted on the UC Irvine machine learning repository.
UR - http://www.scopus.com/inward/record.url?scp=38949172021&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=38949172021&partnerID=8YFLogxK
U2 - 10.1109/MLSP.2006.275514
DO - 10.1109/MLSP.2006.275514
M3 - Conference contribution
SN - 1424406560
SN - 9781424406562
T3 - Proceedings of the 2006 16th IEEE Signal Processing Society Workshop on Machine Learning for Signal Processing, MLSP 2006
SP - 15
EP - 20
BT - Proceedings of the 2006 16th IEEE Signal Processing Society Workshop on Machine Learning for Signal Processing, MLSP 2006
PB - IEEE Computer Society
T2 - 2006 16th IEEE Signal Processing Society Workshop on Machine Learning for Signal Processing, MLSP 2006
Y2 - 6 September 2006 through 8 September 2006
ER -