TY - GEN
T1 - Classifying scientific publications using abstract features
AU - Caragea, Cornelia
AU - Silvescu, Adrian
AU - Kataria, Saurabh
AU - Caragea, Doina
AU - Mitra, Prasenjit
PY - 2011/12/1
Y1 - 2011/12/1
N2 - With the exponential increase in the number of documents available online, e.g., news articles, weblogs, scientific documents, effective and efficient classification methods are required in order to deliver the appropriate information to specific users or groups. The performance of document classifiers critically depends, among other things, on the choice of the feature representation. The commonly used "bag of words" representation can result in a large number of features. Feature abstraction helps reduce a classifier input size by learning an abstraction hierarchy over the set of words. A cut through the hierarchy specifies a compressed model, where the nodes on the cut represent abstract features. In this paper, we compare feature abstraction with two other methods for dimensionality reduction, i.e., feature selection and Latent Dirichlet Allocation (LDA). Experimental results on two data sets of scientific publications show that classifiers trained using abstract features significantly outperform those trained using features that have the highest average mutual information with the class, and those trained using the topic distribution and topic words output by LDA. Furthermore, we propose an approach to automatic identification of a cut in order to trade off the complexity of classifiers against their performance. Our results demonstrate the feasibility of the proposed approach.
AB - With the exponential increase in the number of documents available online, e.g., news articles, weblogs, scientific documents, effective and efficient classification methods are required in order to deliver the appropriate information to specific users or groups. The performance of document classifiers critically depends, among other things, on the choice of the feature representation. The commonly used "bag of words" representation can result in a large number of features. Feature abstraction helps reduce a classifier input size by learning an abstraction hierarchy over the set of words. A cut through the hierarchy specifies a compressed model, where the nodes on the cut represent abstract features. In this paper, we compare feature abstraction with two other methods for dimensionality reduction, i.e., feature selection and Latent Dirichlet Allocation (LDA). Experimental results on two data sets of scientific publications show that classifiers trained using abstract features significantly outperform those trained using features that have the highest average mutual information with the class, and those trained using the topic distribution and topic words output by LDA. Furthermore, we propose an approach to automatic identification of a cut in order to trade off the complexity of classifiers against their performance. Our results demonstrate the feasibility of the proposed approach.
UR - http://www.scopus.com/inward/record.url?scp=84890574323&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84890574323&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:84890574323
SN - 9781577355434
T3 - SARA 2011 - Proceedings of the 9th Symposium on Abstraction, Reformulation, and Approximation
SP - 26
EP - 33
BT - SARA 2011 - Proceedings of the 9th Symposium on Abstraction, Reformulation, and Approximation
T2 - 9th Symposium on Abstraction, Reformulation, and Approximation, SARA 2011
Y2 - 17 July 2011 through 18 July 2011
ER -