The Generalized Training Methods,Algorithms And Applications For Competitive Learning Neural Networks  Posted on:20180807  Degree:Doctor  Type:Dissertation  Country:China  Candidate:Z Y Xiang  Full Text:PDF  GTID:1318330542474511  Subject:Computer Science and Technology  Abstract/Summary:  PDF Full Text Request  The amount of data in the Internet is ever growing because the development of global informatization.Since the information volume is expanding faster than the computational powers that humanity commands,the space complexity of machine learning algorithms is becoming more important than before.Meanwhile,the informatization is spreading into industrial areas other than computer sciences,with the development of the concept and technology of Internet of things and ubiquitous computing.There are a great amount of memory critical computation units in these areas that require low space complex algorithms.Online learning is a machine learning setting,where the machine learning model is acquired by a single pass scan of the training dataset,and it is more space efficient than the traditional offline methods,which need to store the entire dataset for training.Besides,there are various applications,such as network intrusion detection,shortterm electricity requirement prediction and traffic flow prediction,where the training data arrives continuously and the problem scale is growing accordingly.In these applications,the model constructed by machine learning algorithms need to be updated continuously,and online algorithms are a major solution.Though online learning is space efficient,it is difficult to solve nonlinearity,semisupervised learning and model parameter selection all at once,because of the single pass learning setting.Moreover,some online learning algorithms is computationally inefficient in updating the machine learning models.However,the competitive learning has no such problems.Unfortunately,the competitive learning researches mainly focus on clustering.For other tasks such as regression,semisupervised classification and dimensionality reduction,the advantages of nonlinear capability and efficient online learning of competitive learning can not be utilized.This thesis focuses on generalization of the competitive learning algorithms to these tasks.The main contributions are listed as follows.1.A kernel density regression framework is proposed to transform competitive learning clustering to regression algorithms.Online algorithms is either parametric or nonparametric.The parametric algorithms take a unified analytical assumption about the dataset,and it is difficult to select the optimal model with best parameters under the online learning setting.The nonparametric algorithms have no such difficulty,however,the majority of them are unable to give smoothed predictions and suffer from compromised generalization abilities.To address the difficulties both in the parametric and nonparametric learning algorithms in online learning,this thesis proposes a nonparametric training method to construct a parametric model.By exploiting the distribution learning abilities of competitive learning neural networks,the joint distribution of explanatory variables and response variables can be modeled,and the regression model be obtained by density regression.Then,the regression function is deducted directly from clustering learning results in this thesis.The complete framework is a feedforward neural network model,the weights of the hidden layer of which is trained by the competitive learning neural networks.The advantage of this framework is that the local minima of the back propagation is avoided,while at the same time the smoothed prediction advantage of feedforward neural network is kept.The proposed methods has constant space complexity and O(n)computation complexity.Experiments on six UCI datasets show that the proposed method gives comparable results to the mainstream offline methods.2.The optimal smooth parameter selection of kernel density regression is deducted by maximum likelihood estimation.There is one smooth parameter in the kernel density regression framework,selection of which is difficult to reach optima in online learning,since the learning is finished in one pass scan.To solve this problem,we propose a maximum likelihood smooth parameter selection framework.Under the online learning setting,the optimal smooth parameter selection of kernel density regression is solved by two strategies of maximum likelihood estimation.To solve the parameter selection of mixture models,a Bayesian inferences process is employed to transform the maximum likelihood function into solvable form.Then through derivation techniques,equations about the smooth parameter are constructed and solved.Besides,a semisupervised learning process is also proposed to select the smooth parameters with the complete distribution information of the test dataset.The proposed framework selection is the optimal solution under the online learning setting,which is vital for the online kernel density regression,because the online algorithms is finished within one pass scan of data and leaves no room for parameter tuning.Experiments show that the proposed method is comparable to other online and offline algorithms,and it is parameter free.3.An inverse competitive learning paradigm is proposed for semisupervised extensions of competitive learning neural networks.The original competitive learning neural networks,especially the selforganizing neural networks are inefficient to handle the labeled datasets,which make them difficult to solve the semisupervised learning problem,where the training dataset contains both labeled and unlabeled samples.To solve this problem,this thesis proposes the inverse competitive learning.Whenever there is a conflicting of labels,the neurons are moved in an inverse direction from the original competitive learning.Then the graph cut algorithm is employed to complete the semisupervised learning.Moreover,a fast graph cut algorithm based on greedy label propagation is proposed.The main advantage of this framework is the robustness to noise in datasets and the ability to perform online semisupervised learning under the manifold assumption.The experiments on the intrusion detection dataset show that the proposed method is more accurate than the mainstream offline semisupervised learning methods even that it is online.4.A framework is proposed to transform clustering into dimensionality reduction.Competitive learning is used in combine for online dimensionality reduction learning.The online learning and nonlinear dimensionality reduction are always difficult to combine.With competitive learning at the core,this thesis proposes an online nonlinear dimensionality reduction framework.First,a data prototype is acquired by the semisupervised competitive learning neural networks.Second,two similarity matrixes are fusioned into one by a quadratic programming process.Finally,the neurons are transformed into the lower dimensional space and other samples are embedded by a kernel smoothing procedure.Besides,a data visualization techniques is proposed exploiting the topology learning abilities of competitive learning.The proposed framework has the advantage of high space complexity comparing to existing nonlinear algorithms.Experimental results on the intrusion detection dataset show that the proposed method is a more efficient alternative to the offline semisupervised learning algorithms.5.Shortterm traffic flow prediction is implemented based on the ensemble of the proposed kernel density regression.Since the prediction results and weights of the individual online regressors are changing constantly with the time passage,the ensemble of such online regressors are always difficult.To address this difficulty,this thesis proposes a weight storing technique to construct a competitive regressor ensemble framework.By construction of weighted learning,the ensemble of kernel density regression is enabled.The proposed framework is a combination of online learning and ensemble learning.Since the ensemble enables the combined use of competitive learning based regression and other algorithms,it is vital for making the proposed regression algorithms applicable in the various industrial and social applications.In the experiments,the proposed method is boosted by Adaboost and the results show that the proposed methods outperform mainstream methods such as support vector machine,decision trees and extreme learning machines etc.,and they have the parameter free advantage.  Keywords/Search Tags:  Online learning, competitive learning, nonparametric regression, semisupervised learning, semisupervised dimensionality reduction  PDF Full Text Request  Related items 
 
