Best Practices of Feature Selection in Multi-Omics Data

Best Practices of Feature Selection in Multi-Omics Data

Funda Ipekten, Gözde Ertürk Zararsız, Halef Okan Doğan, Vahap Eldem, Gökmen Zararsız
DOI: 10.4018/979-8-3693-3026-5.ch014
OnDemand:
(Individual Chapters)
Available
$33.75
List Price: $37.50
10% Discount:-$3.75
TOTAL SAVINGS: $3.75

Abstract

With the recent advances in molecular biology techniques such as next-generation sequencing, mass-spectrometry, etc., a large omic data is produced. Using such data, the expression levels of thousands of molecular features (genes, proteins, metabolites, etc.) can be quantified and associated with diseases. The fact that multiple omics data contains different types of data and the number of analyzed variables increases the complexity of the models created with machine learning methods. In addition, due to many variables, the investigation of molecular variables associated with diseases is very costly. Therefore, selecting the informative and disease-related molecular features is applicable before model training and evaluation. This feature selection step is essential for obtaining accurate and generalizable models in minimum time with minimum cost. Some current methods used for feature selection are as follows: recursive feature elimination, information gain, minimum redundancy maximum relevance (mRMR), boruta, altmann, and lasso.
Chapter Preview

Previously published in the Encyclopedia of Data Science and Machine Learning; pages 2045-2059, copyright year 2023 by Engineering Science Reference (an imprint of IGI Global).

Top

Introduction

Today, there is an increase in data in many areas. With this increase, the number and variety of the variables to be evaluated also increases. The increase in data and variables became a situation that needed to be solved among world problems. In addition, although there is a perception that having too much data in the scientific field, having too much information, correct information, or sufficient information may not be possible. However, it should not be forgotten that there is valuable information in a relatively large amount of data. It should be clear that it can be beneficial to have much data to extract this helpful information. However, performing data analyses to obtain and process this information can be difficult. In addition, its existence is a problem called the curse of data dimensionality (Verkeysen M. and François D., 2005). High-dimensional data sets, where these problems are most common, are used successfully in multiple fields such as genetics, pharmacology, toxicology, nutrition, and genetics. The use of these high-dimensional data allows one to examine biology systems, cellular metabolism, and disease etiologies in more detail. However, the number of samples (n) of these data is considerably lower than the number of variables (p) and the heterogeneity of the data, the missing observations in the data as a result of the use of high-output technology, limits the use of traditional methods that can be used in this field. Therefore, there is a need for the clinical understanding of the biological system based on research and machine learning, and statistical learning methods to analyze this clinical information statistically (Hastie et al., 2009). Several studies are show that machine learning methods are used and applied successfully in studies carried out in this field. Some of these studies are listed in Table 1.

Table 1.
Some studies using feature selection
DatasetsMethodsReferences
Ovarian cancerClassificationMKLWilson et al.2019
Breast cancerClassificationMKLTao et al.2019
Gene expressionClassificationSVMGolub et al.1999
Gut microbiotaClassificationRFFranzosa et al.2019
Colon cancerClassificationSVMMoler et al.2000
Ovarian, leukemia, colonClassificationSVMFurey et al.2000

MKL: Multiple Kernel Learning, SVM: Support Vector Machine, RF:Random Forest

Complete Chapter List

Search this Book:
Reset