High-Dimensional Feature Selection for Genomic Datasets

02/27/2020
by   Majid Afshar, et al.
0

In the presence of large dimensional datasets that contain many irrelevant features (variables), dimensionality reduction algorithms have proven to be useful in removing features with low variance and combine features with high correlation. In this paper, we propose a new feature selection method which uses singular value decomposition of a matrix and the method of least squares to remove the irrelevant features and detect correlations between the remaining features. The effectiveness of our method has been verified by performing a series of comparisons with state-of-the-art feature selection methods over ten genetic datasets ranging up from 9,117 to 267,604 features. The results show that our method is favorable in various aspects compared to state-of-the-art feature selection methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/26/2019

A Feature Selection Based on Perturbation Theory

Consider a supervised dataset D=[A|b], where b is the outcome column, ro...
research
11/20/2018

Contingency Training

When applied to high-dimensional datasets, feature selection algorithms ...
research
06/06/2017

Embedding Feature Selection for Large-scale Hierarchical Classification

Large-scale Hierarchical Classification (HC) involves datasets consistin...
research
12/17/2022

An Evolutionary Multitasking Algorithm with Multiple Filtering for High-Dimensional Feature Selection

Recently, evolutionary multitasking (EMT) has been successfully used in ...
research
10/22/2019

Orthogonal variance decomposition based feature selection

Existing feature selection methods fail to properly account for interact...
research
10/05/2019

Covariance-free Partial Least Squares: An Incremental Dimensionality Reduction Method

Dimensionality reduction plays an important role in computer vision prob...

Please sign up or login with your details

Forgot password? Click here to reset