Font Size: a A A

Provable and Practical Algorithms for Non-Convex Problems in Machine Learnin

Posted on:2019-10-09Degree:Ph.DType:Dissertation
University:Cornell UniversityCandidate:Yuan, YangFull Text:PDF
GTID:1478390017485083Subject:Computer Science
Abstract/Summary:
Machine learning has become one of the most exciting research areas in the world, with various applications. However, there exists a noticeable gap between theory and practice. On one hand, a simple algorithm like stochastic gradient descent (SGD) works very well in practice, without satisfactory theoretical explanations. On the other hand, the algorithms analyzed in the theoretical machine learning literature, although with solid guarantees, tend to be less efficient compared with the techniques widely used in practice, which are usually hand tuned or ad hoc based on intuition.;This dissertation is about bridging the gap between theory and practice from two directions. The first direction is "practice to theory", i.e., to explain and analyze the existing algorithms and empirical observations in machine learning. Along this direction, we provide sufficient conditions for SGD to escape saddle points and local minima, as well as SGD dynamics analysis for the two-layer neural network with ReLU activation.;The other direction is "theory to practice", i.e., using theoretical tools to obtain new, better and practical algorithms. Along this direction, we introduce a new algorithm Harmonica that uses Fourier analysis and compressed sensing for tuning hyperparameters. Harmonica supports parallel sampling and works well for tuning neural networks with more than 30 hyperparameters.
Keywords/Search Tags:Machine, Algorithms
Related items