Font Size: a A A

Multi-scale adaptive representation of signals: Models and algorithms

Posted on:2017-02-21Degree:Ph.DType:Thesis
University:Princeton UniversityCandidate:Tai, ChengFull Text:PDF
GTID:2468390014455371Subject:Applied Mathematics
Abstract/Summary:
Representations of data play a key role in many signal processing and machine learning applications. For low-level signal processing tasks, dictionary learning is a very popular approach for representing signals and has been successfully used in image/video denoising, compression and inpainting tasks. For high-level signal processing tasks such as object recognition, deep learning models, especially convolutional neural networks (CNN), are becoming an increasingly popular approach to provide abstract representations of signals. Dictionary learning and deep learning are two motivating sources of this thesis.;In this thesis, we study the modeling and algorithmic aspects of building multi-scale adaptive representations of signals. Special attention is given to the computational efficiency of such representations.;In the first part of this thesis, we provide a framework for constructing adaptive waveletframes and bi-frames (abbreviated as AdaFrame). This framework gives multi-scale, sparse representations of the signal, with an efficiency comparable to that of the wavelets at inference time. Similar to dictionary learning, it is also adapted to data. These features make AdaFrame an attractive alternative to dictionary learning and the more conventional wavelet frames. The proposed framework is formally similar to the first few layers of a convolutional network. As a byproduct, we show that the proposed framework gives a better way of visualizing the activations of the intermediate layers of a neural net in terms of reconstruction error. Some examples are given to demonstrate the wide applicability of AdaFrame, including image denoising, image compression, object recognition and video super-resolution.;In the second part of the this thesis, we consider accelerating CNNs by low-rank approximations. CNNs are typical examples of adaptive representations of signals, and have delivered impressive performance in various computer vision applications. However the storage and computation requirements make it problematic for deploying these models on mobile devices. We propose a new tensor decomposition technique for accelerating CNNs. As a result, typical modern CNNs can be accelerated by a factor up to 3 and the number of parameters is also reduced by a significant factor.;In the third part of the thesis, we consider the dynamics of stochastic gradient algorithms (SGAs). SGAs have become the algorithm of choice for large scale machine learning applications. They are used for training AdaFrame and CNNs on large datasets. As we focus on the computational efficiency, we want to understand the dynamics of SGAs as well as ways to improve them. We propose the method of stochastic modified equations (SME) to analyze the dynamics of the SGA. Using this technique, we can give precise characterizations for both the initial convergence speed and the eventual oscillations, at least in some special cases. Furthermore, the SME formalism allows us to characterize various speed-up techniques, such as introducing momentum, adjusting the learning rate and adjusting the mini-batch sizes. Previously, these techniques relied mostly on heuristics. Besides introducing simple examples to illustrate the SME formalism, we also apply the framework to improve the relaxed randomized Kaczmarz method for solving linear equations. The SME framework is a precise and unifying approach to understanding and improving the SGA, and has the potential to be applied to many more stochastic algorithms.
Keywords/Search Tags:Signal, Adaptive, Representations, Multi-scale, Models, Dictionary learning, SME
Related items