On Interacting Features in Subset Selection

On Interacting Features in Subset Selection

Zheng Zhao
Copyright: © 2009 |Pages: 6
ISBN13: 9781605660103|ISBN10: 1605660108|EISBN13: 9781605660110
DOI: 10.4018/978-1-60566-010-3.ch167
Cite Chapter Cite Chapter

MLA

Zhao, Zheng. "On Interacting Features in Subset Selection." Encyclopedia of Data Warehousing and Mining, Second Edition, edited by John Wang, IGI Global, 2009, pp. 1079-1084. https://doi.org/10.4018/978-1-60566-010-3.ch167

APA

Zhao, Z. (2009). On Interacting Features in Subset Selection. In J. Wang (Ed.), Encyclopedia of Data Warehousing and Mining, Second Edition (pp. 1079-1084). IGI Global. https://doi.org/10.4018/978-1-60566-010-3.ch167

Chicago

Zhao, Zheng. "On Interacting Features in Subset Selection." In Encyclopedia of Data Warehousing and Mining, Second Edition, edited by John Wang, 1079-1084. Hershey, PA: IGI Global, 2009. https://doi.org/10.4018/978-1-60566-010-3.ch167

Export Reference

Mendeley
Favorite

Abstract

The high dimensionality of data poses a challenge to learning tasks such as classification. In the presence of many irrelevant features, classification algorithms tend to overfit training data (Guyon & Elisseeff, 2003). Many features can be removed without performance deterioration, and feature selection is one effective means to remove irrelevant features (Liu & Yu, 2005). Feature selection, also known as variable selection, feature reduction, attribute selection or variable subset selection, is the technique of selecting a subset of relevant features for building robust learning models. Usually a feature is relevant due to two reasons: (1) it is strongly correlated with the target concept; or (2) it forms a feature subset with other features and the subset is strongly correlated with the target concept. Optimal feature selection requires an exponentially large search space (O(2n), where n is the number of features) (Almual-lim & Dietterich, 1994). Researchers often resort to various approximations to determine relevant features, and in many existing feature selection algorithms, feature relevance is determined by correlation between individual features and the class (Hall, 2000; Yu & Liu, 2003). However, a single feature can be considered irrelevant based on its correlation with the class; but when combined with other features, it can become very relevant. Unintentional removal of these features can result in the loss of useful information and thus may cause poor classification performance, which is studied as attribute interaction in (Jakulin & Bratko, 2003). Therefore, it is desirable to consider the effect of feature interaction in feature selection.

Request Access

You do not own this content. Please login to recommend this title to your institution's librarian or purchase it from the IGI Global bookstore.