TY - JOUR
T1 - Enhancing training collections for image annotation
T2 - An instance-weighted mixture modeling approach
AU - Sawant, Neela
AU - Wang, James Z.
AU - Li, Jia
PY - 2013/9
Y1 - 2013/9
N2 - Tagged Web images provide an abundance of labeled training examples for visual concept learning. However, the performance of automatic training data selection is susceptible to highly inaccurate tags and atypical images. Consequently, manually curated training data sets are still a preferred choice for many image annotation systems. This paper introduces ARTEMIS - a scheme to enhance automatic selection of training images using an instance-weighted mixture modeling framework. An optimization algorithm is derived to learn instance-weights in addition to mixture parameter estimation, essentially adapting to the noise associated with each example. The mechanism of hypothetical local mapping is evoked so that data in diverse mathematical forms or modalities can be cohesively treated as the system maintains tractability in optimization. Finally, training examples are selected from top-ranked images of a likelihood-based image ranking. Experiments indicate that ARTEMIS exhibits higher resilience to noise than several baselines for large training data collection. The performance of ARTEMIS-trained image annotation system is comparable with usage of manually curated data sets.
AB - Tagged Web images provide an abundance of labeled training examples for visual concept learning. However, the performance of automatic training data selection is susceptible to highly inaccurate tags and atypical images. Consequently, manually curated training data sets are still a preferred choice for many image annotation systems. This paper introduces ARTEMIS - a scheme to enhance automatic selection of training images using an instance-weighted mixture modeling framework. An optimization algorithm is derived to learn instance-weights in addition to mixture parameter estimation, essentially adapting to the noise associated with each example. The mechanism of hypothetical local mapping is evoked so that data in diverse mathematical forms or modalities can be cohesively treated as the system maintains tractability in optimization. Finally, training examples are selected from top-ranked images of a likelihood-based image ranking. Experiments indicate that ARTEMIS exhibits higher resilience to noise than several baselines for large training data collection. The performance of ARTEMIS-trained image annotation system is comparable with usage of manually curated data sets.
UR - http://www.scopus.com/inward/record.url?scp=84893612413&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84893612413&partnerID=8YFLogxK
U2 - 10.1109/TIP.2013.2262289
DO - 10.1109/TIP.2013.2262289
M3 - Article
C2 - 23674453
AN - SCOPUS:84893612413
SN - 1057-7149
VL - 22
SP - 3562
EP - 3577
JO - IEEE Transactions on Image Processing
JF - IEEE Transactions on Image Processing
IS - 9
ER -