Abstract
The paper presents an original filter approach for effective feature selection in classification tasks with a very large number of input variables. The approach is based on the use of a new information theoretic selection criterion: the double input symmetrical relevance (DISR). The rationale of the criterion is that a set of variables can return an information on the output class that is higher than the sum of the informations of each variable taken individually. This property will be made explicit by defining the measure of variable complementarity. A feature selection filter based on the DISR criterion is compared in theoretical and experimental terms to recently proposed information theoretic criteria. Experimental results on a set of eleven microarray classification tasks show that the proposed technique is competitive with existing filter selection methods.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. Journal of Machine Learning Research 3, 1157–1182 (2003)
Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artificial Intelligence 97, 273–324 (1997)
Blum, A., Langley, P.: Selection of relevant features and examples in machine learning. Artificial Intelligence 97, 245–271 (1997)
Provan, G., Singh, M.: Learning bayesian networks using feature selection. In: Fifth International Workshop on Artificial Intelligence and Statistics, pp. 450–456 (1995)
Duch, W., Winiarski, T., Biesiada, J., Kachel, A.: Feature selection and ranking filters. In: Kaynak, O., Alpaydın, E., Oja, E., Xu, L. (eds.) ICANN 2003 and ICONIP 2003. LNCS, vol. 2714, pp. 251–254. Springer, Heidelberg (2003)
Bell, D.A., Wang, H.: A formalism for relevance and its application in feature subset selection. Machine Learning 41, 175–195 (2000)
Peng, H., Long, F.: An efficient max-dependency algorithm for gene selection. In: 36th Symposium on the Interface: Computational Biology and Bioinformatics (2004)
Fleuret, F.: Fast binary feature selection with conditional mutual information. Journal of Machine Learning Research 5, 1531–1555 (2004)
Yu, L., Liu, H.: Efficient feature selection via analysis of relevance and redundancy. Journal of Machine Learning Research 5, 1205–1224 (2004)
Cover, T.M., Thomas, J.A.: Elements of Information Theory. John Wiley, New York (1990)
Yang, H., Moody, J.: Feature selection based on joint mutual information. In: Advances in Intelligent Data Analysis (AIDA), Computational Intelligence Methods and Applications (CIMA), Rochester New York, ICSC (1999)
Kojadinovic, I.: Relevance measures for subset variable selection in regression problems based on k-additive mutual information. Computational Statistics and Data Analysis 49, 1205–1227 (2005)
Meyer, P.: Information theoretic filters for feature selection. Technical report, Universite Libre de Bruxelles (548) (2005)
web, http://www.tech.plym.ac.uk/spmc/bioinformatics/microarraycancers.html
Scott, D.W.: Multivariate Density Estimation. Wiley, Chichester (1992)
R-project, www.r-project.org
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Meyer, P.E., Bontempi, G. (2006). On the Use of Variable Complementarity for Feature Selection in Cancer Classification. In: Rothlauf, F., et al. Applications of Evolutionary Computing. EvoWorkshops 2006. Lecture Notes in Computer Science, vol 3907. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11732242_9
Download citation
DOI: https://doi.org/10.1007/11732242_9
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-33237-4
Online ISBN: 978-3-540-33238-1
eBook Packages: Computer ScienceComputer Science (R0)Springer Nature Proceedings Computer Science
Keywords
- Feature Selection
- Mutual Information
- Versus Versus Versus Versus
- Feature Selection Algorithm
- Versus Versus Versus Versus Versus
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.