TY - GEN
T1 - Improved fine-grained component-conditional class labeling with active learning
AU - Miller, David J.
AU - Lin, Chu Fang
AU - Kesidis, George
AU - Collins, Christopher M.
PY - 2010
Y1 - 2010
N2 - We have recently introduced new generative semisupervised mixtures with more fine-grained class label generation mechanisms than previous methods [11], [12]. Our models combine advantages of semisupervised mixtures, which achieve label extrapolation over a component, and nearest-neighbor (NN)/nearest- prototype (NP) classification, which achieves accurate classification in the vicinity of labeled samples. Our models are advantageous when within-component class proportions are not constant over the feature space region "owned by" a component. In this paper, we develop an active learning extension of our fine-grained labeling methods. We propose two new uncertainty sampling methods in comparison with traditional entropy-based uncertainty sampling. Our experiments on a number of UC Irvine data sets show that the proposed active learning methods improve classification accuracy more than standard entropybased active learning. The proposed methods are particularly advantageous when the labeled percentage is small. We also extend our semisupervised method to allow variable weighting on labeled and unlabeled data likelihood terms. This approach is shown to outperform previous weighting schemes.
AB - We have recently introduced new generative semisupervised mixtures with more fine-grained class label generation mechanisms than previous methods [11], [12]. Our models combine advantages of semisupervised mixtures, which achieve label extrapolation over a component, and nearest-neighbor (NN)/nearest- prototype (NP) classification, which achieves accurate classification in the vicinity of labeled samples. Our models are advantageous when within-component class proportions are not constant over the feature space region "owned by" a component. In this paper, we develop an active learning extension of our fine-grained labeling methods. We propose two new uncertainty sampling methods in comparison with traditional entropy-based uncertainty sampling. Our experiments on a number of UC Irvine data sets show that the proposed active learning methods improve classification accuracy more than standard entropybased active learning. The proposed methods are particularly advantageous when the labeled percentage is small. We also extend our semisupervised method to allow variable weighting on labeled and unlabeled data likelihood terms. This approach is shown to outperform previous weighting schemes.
UR - https://www.scopus.com/pages/publications/79952434390
UR - https://www.scopus.com/pages/publications/79952434390#tab=citedBy
U2 - 10.1109/ICMLA.2010.8
DO - 10.1109/ICMLA.2010.8
M3 - Conference contribution
AN - SCOPUS:79952434390
SN - 9780769543000
T3 - Proceedings - 9th International Conference on Machine Learning and Applications, ICMLA 2010
SP - 3
EP - 8
BT - Proceedings - 9th International Conference on Machine Learning and Applications, ICMLA 2010
T2 - 9th International Conference on Machine Learning and Applications, ICMLA 2010
Y2 - 12 December 2010 through 14 December 2010
ER -