Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01sf2687865
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorLiu, Han-
dc.contributor.authorWang, Zhaoran-
dc.contributor.otherOperations Research and Financial Engineering Department-
dc.date.accessioned2018-10-09T21:09:39Z-
dc.date.available2018-10-09T21:09:39Z-
dc.date.issued2018-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01sf2687865-
dc.description.abstractComplex statistical models, combined with scalable computing architectures, are beginning to shed lights on the analysis of large, intricate, and noisy datasets. To glue these three components --- big data, big models, and big architectures --- we need a seamless integration of high-dimensional statistics and large-scale optimization. In this context, the high-level goal of this thesis is to develop a new generation of {\it statistical optimization} method and theory to address emerging challenges in data science and artificial intelligence. In particular, {\it nonconvex} formulations of machine learning problems bring superior computational efficiency, statistical accuracy, and modeling flexibility. However, a gap separates theory from practice. On the one hand, nonconvex optimization heuristics often exhibit good empirical performance in practice. On the other hand, classical theory only characterizes the statistical accuracy of a hypothetical global optimum, which is computationally intractable to attain in the worst case. Such a gap between theory and practice prohibits us from designing efficient algorithms in a principled way. This thesis aims to answer the question: How should we develop nonconvex statistical optimization algorithms with provable guarantees? In this thesis, we develop a new unified algorithmic framework that integrates the {\it global exploration} and {\it local exploitation} of the {\it latent convexity} induced by the randomness of data. Under such a framework, we propose two global exploration schemes, namely, {\it homotopy path following} and {\it tightening after relaxation}, in the two parts of this thesis, respectively. Combining the two global exploration schemes with model-specific local exploitation methods, we establish globally convergent and statistically optimal nonconvex statistical optimization algorithms correspondingly for two problems, namely, nonconvex $M$-estimation and sparse principal component analysis.-
dc.language.isoen-
dc.publisherPrinceton, NJ : Princeton University-
dc.relation.isformatofThe Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the library's main catalog: <a href=http://catalog.princeton.edu> catalog.princeton.edu </a>-
dc.subjectmachine learning-
dc.subjectoptimization-
dc.subjectstatistics-
dc.subject.classificationOperations research-
dc.subject.classificationComputer science-
dc.subject.classificationElectrical engineering-
dc.titleNonconvex Statistical Optimization-
dc.typeAcademic dissertations (Ph.D.)-
pu.projectgrantnumber690-2143-
Appears in Collections:Operations Research and Financial Engineering

Files in This Item:
File Description SizeFormat 
Wang_princeton_0181D_12672.pdf1.81 MBAdobe PDFView/Download


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.