Abstract
Feature selection is used to improve the efficiency of learning algorithms by finding an optimal subset of features. However, most feature selection techniques can handle only certain types of data. Additional limitations of existing methods include intensive computational requirements and inability to identify redundant variables. In this paper, we present a novel, information-theoretic algorithm for feature selection, which finds an optimal set of attributes by removing both irrelevant and redundant features. The algorithm has a polynomial computational complexity and is applicable to datasets of a mixed nature. The method performance is evaluated on several benchmark datasets by using a standard classifier (C4.5).
Original language | English |
---|---|
Pages (from-to) | 799-811 |
Number of pages | 13 |
Journal | Pattern Recognition Letters |
Volume | 22 |
Issue number | 6-7 |
DOIs | |
State | Published - 1 May 2001 |
Externally published | Yes |
Keywords
- Classification
- Feature selection
- Information theory
- Information-theoretic network
ASJC Scopus subject areas
- Software
- Signal Processing
- Computer Vision and Pattern Recognition
- Artificial Intelligence