Please use this identifier to cite or link to this item:
http://arks.princeton.edu/ark:/88435/dsp019k41zd62n
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Arora, Sanjeev | en_US |
dc.contributor.author | Ge, Rong | en_US |
dc.contributor.other | Computer Science Department | en_US |
dc.date.accessioned | 2013-12-06T14:16:33Z | - |
dc.date.available | 2013-12-06T14:16:33Z | - |
dc.date.issued | 2013 | en_US |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/dsp019k41zd62n | - |
dc.description.abstract | Modern machine learning algorithms can extract useful information from text, images and videos. All these applications involve solving NP-hard problems in average case using heuristics. What properties of the input allow it to be solved efficiently? Theoretically analyzing the heuristics is very challenging. Few results were known. This thesis takes a different approach: we identify natural properties of the input, then design new algorithms that provably works assuming the input has these properties. We are able to give new, provable and sometimes practical algorithms for learning tasks related to text corpus, images and social networks. The first part of the thesis presents new algorithms for learning thematic structure in documents. We show under a reasonable assumption, it is possible to provably learn many topic models, including the famous Latent Dirichlet Allocation. Our algorithm is the first provable algorithms for topic modeling. An implementation runs 50 times faster than latest MCMC implementation and produces comparable results. The second part of the thesis provides ideas for provably learning deep, sparse representations. We start with sparse linear representations, and give the first algorithm for dictionary learning problem with provable guarantees. Then we apply similar ideas to deep learning: under reasonable assumptions our algorithms can learn a deep network built by denoising autoencoders. The final part of the thesis develops a framework for learning latent variable models. We demonstrate how various latent variable models can be reduced to orthogonal tensor decomposition, and then be solved using tensor power method. We give a tight sample complexity analysis for tensor power method, which reduces the number of sample required for learning many latent variable models. In theory, the assumptions in this thesis help us understand why intractable problems in machine learning can often be solved; in practice, the results suggest inherently new approaches for machine learning. We hope the assumptions and algorithms inspire new research problems and learning algorithms. | en_US |
dc.language.iso | en | en_US |
dc.publisher | Princeton, NJ : Princeton University | en_US |
dc.relation.isformatof | The Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the <a href=http://catalog.princeton.edu> library's main catalog </a> | en_US |
dc.subject | machine learning | en_US |
dc.subject | matrix factorization | en_US |
dc.subject | provable algorithms | en_US |
dc.subject | tensor decomposition | en_US |
dc.subject | topic models | en_US |
dc.subject.classification | Computer science | en_US |
dc.title | Provable Algorithms for Machine Learning Problems | en_US |
dc.type | Academic dissertations (Ph.D.) | en_US |
pu.projectgrantnumber | 690-2143 | en_US |
Appears in Collections: | Computer Science |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Ge_princeton_0181D_10819.pdf | 2.15 MB | Adobe PDF | View/Download |
Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.