Font Size: a A A

Research On Sparse Learning Based Multiple Indefinite Kernel Learning For Feature Selection Algorithms

Posted on:2020-10-06Degree:MasterType:Thesis
Country:ChinaCandidate:Y SongFull Text:PDF
GTID:2428330623459881Subject:Software engineering
Abstract/Summary:PDF Full Text Request
Feature selection refers to the process of selecting some of the most effective features from the original features sets to reduce the dimensions of the datasets.It can reduce the complexity of the model and the risk of overfitting.In recent years,there are many researches on feature selection algorithms.Multiple kernel learning for feature selection?MKL-FS?utilizes kernels to explore complex properties of features and performs better in linear and nonlinear feature selection algorithms.However,MKL-FS has two limitations:?1?the choice of kernel function is not rich enough;?2?the solution of feature selection is not sparse enough.On the one hand,MKL-FS requires the kernel function to satisfy the positive definite condition,which constrains the rich expression ability of the kernel function to some extent.Recent research shows that indefinite kernels can better characterize the relationship between data and achieve better results than positive definite kernels in many practical applications.However,due to the non-convexity of indefinite kernels,existing MKL-FS methods are usually inapplicable and the corresponding research is also relatively little.On the other hand,previous MKL-FS methods usually use the l1-norm to obtain sparse kernel combination coefficients.However,the l1-norm,as a convex approximation of l0-norm,sometimes cannot attain the desired solution of the l0-norm regularizer problem and may lead to prediction accuracy loss.Since the optimization problem of the l0-norm is NP-hard,many linear feature selection methods use various non-convex approximations of l0-norm to replace l0-norm,and have achieved good results.However,non-convex approximations are rarely used in nonlinear models currently.This paper will improve the MKL-FS methods from these two aspects and the work are summarized as follows:1)Since MKL-FS methods are limited to using positive definite kernels,this paper proposes a novel l1-norm based multiple indefinite kernel learning for feature selection?l1-MIK? method based on the primal framework of indefinite kernel support vector machine?IKSVM?,which applies an indefinite base kernel for each feature and then exerts an l1-norm constraint on kernel combination coefficients to select features automatically.In order to solve the non-convex optimization model,a two-stage algorithm is proposed to optimize the IKSVM coefficients and the kernel combination coefficients alternatively.In the algorithm,the non-convex problem of the IKSVM is reconstructed into a difference of convex functions?DC? programming,and solved by DC algorithms.The optimization problem of the kernel combination coefficient is solved by the projected gradient method.In order to further extend to large-scale problems,this paper uses a leverage score method to sample large-scale datasets,and extends the l1-MIK method to multi-class classification scenarios.Finally,the effectiveness of the proposed algorithm is verified on the actual datasets.2)Since MKL-FS methods are limited to using l1-norm for feature selection,this paper further proposes a novel l0-norm based multiple indefinite kernel?l0-MIK?method for feature selection with non-convex approximations constraint on kernel combination coefficients to select features automatically.l0-MIK is based on the primal framework of IKSVM and is also solved by a two-stage algorithm,in which the non-convex problem of the IKSVM and the non-convex problem of the l0-norm are reconstructed into DC programming,and solved by DC algorithms respectively.Extensive experiments show that l0-MIK is superior to the existing MKL-FS methods and l0-norm based linear feature selection algorithms in terms of classification classification accuracy and sparsity of features.
Keywords/Search Tags:Multiple indefinite kernel learning, l0-norm, Feature selection, Difference of convex functions programming
PDF Full Text Request
Related items