Evolving feature selection
- 12 December 2005
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Intelligent Systems
- Vol. 20 (6), 64-76
- https://doi.org/10.1109/mis.2005.105
Abstract
Data preprocessing is an indispensable step in effective data analysis. It prepares data for data mining and machine learning, which aim to turn data into business intelligence or knowledge. Feature selection is a preprocessing technique commonly used on high-dimensional data. Feature selection studies how to select a subset or list of attributes or variables that are used to construct models describing data. Its purposes include reducing dimensionality, removing irrelevant and redundant features, reducing the amount of data needed for learning, improving algorithms' predictive accuracy, and increasing the constructed models' comprehensibility. This article considers feature-selection overfitting with small-sample classifier design; feature selection for unlabeled data; variable selection using ensemble methods; minimum redundancy-maximum relevance feature selection; and biological relevance in feature selection for microarray data.Keywords
This publication has 5 references indexed in Scilit:
- Impact of error estimation on feature selectionPattern Recognition, 2005
- Grand challenges for multimodal bio-medical systemsIEEE Circuits and Systems Magazine, 2005
- Genomic signal processing: diagnosis and therapyIEEE Signal Processing Magazine, 2005
- Optimal number of features as a function of sample size for various classification rulesBioinformatics, 2004
- On the Possible Orderings in the Measurement Selection ProblemIEEE Transactions on Systems, Man, and Cybernetics, 1977