Feature selection and classification using ensembles of genetic programs and within-class and between-class permutations
2015 (English)In: 2015 IEEE Congress on Evolutionary Computation, CEC 2015, IEEE , 2015, 1121-1128 p.Conference paper (Refereed)Text
Many feature selection methods are based on the assumption that important features are highly correlated with their corresponding classes, but mainly uncorrelated with each other. Often, this assumption can help eliminate redundancies and produce good predictors using only a small subset of features. However, when the predictability depends on interactions between features, such methods will fail to produce satisfactory results. In this paper a method that can find important features, both independently and dependently discriminative, is introduced. This method works by performing two different types of permutation tests that classify each of the features as either irrelevant, independently predictive or dependently predictive. It was evaluated using a classifier based on an ensemble of genetic programs. The attributes chosen by the permutation tests were shown to yield classifiers at least as good as the ones obtained when all attributes were used during training-and often better. The proposed method also fared well when compared to other attribute selection methods such as RELIEFF and CFS. Furthermore, the ability to determine whether an attribute was independently or dependently predictive was confirmed using artificial datasets with known dependencies.
Place, publisher, year, edition, pages
IEEE , 2015. 1121-1128 p.
Probability Theory and Statistics
IdentifiersURN: urn:nbn:se:kth:diva-187112DOI: 10.1109/CEC.2015.7257015ScopusID: 2-s2.0-84963617435ISBN: 978-147997492-4OAI: oai:DiVA.org:kth-187112DiVA: diva2:929666
IEEE Congress on Evolutionary Computation, CEC 2015; Sendai; Japan
QC 201605192016-05-192016-05-172016-05-19Bibliographically approved