Please use this identifier to cite or link to this item:
http://arks.princeton.edu/ark:/88435/dsp012z10wt107
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Liu, Han | - |
dc.contributor.author | Eisenach, Carson | - |
dc.contributor.other | Operations Research and Financial Engineering Department | - |
dc.date.accessioned | 2019-11-05T16:47:22Z | - |
dc.date.available | 2019-11-05T16:47:22Z | - |
dc.date.issued | 2019 | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/dsp012z10wt107 | - |
dc.description.abstract | Traditional problems in statistics and machine learning are relatively well understood -- they often feature low dimensionality, convex loss functions, and independent, identically distributed data. By contrast, many modern learning problems feature high dimensional data, non-convex learning objectives, and data distributions that change during the learning process. Whether the problem of interest is labeled as statistics, machine learning, statistical learning, or reinforcement learning, methods for solving it can be viewed as the stochastic optimization of some objective function. Accordingly, we address the aforementioned challenges via the lens of statistical optimization -- a statistical approach for understanding and solving stochastic optimization. In particular, we focus on deriving new methodology with computational and statistical guarantees for two classes of problems: recovering and performing inference on latent patterns in high-dimensional graphical models, and continuous control over bounded action spaces. In the first part of this dissertation, we consider a class of cluster-based graphical models. We introduce a novel algorithm for variable clustering named FORCE, based on solving a convex relaxation of the K-means criterion, as well as post-dimension reduction inferential procedures. In the second part, we consider the reinforcement learning (RL) setting, where an agent seeks to learn a decision-making policy based on feedback from its environment. We derive a novel class of variance-reduced estimators called Marginal Policy Gradients, and demonstrate both their improved statistical properties and their application to several control tasks. | - |
dc.language.iso | en | - |
dc.publisher | Princeton, NJ : Princeton University | - |
dc.relation.isformatof | The Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the library's main catalog: <a href=http://catalog.princeton.edu> catalog.princeton.edu </a> | - |
dc.subject.classification | Statistics | - |
dc.subject.classification | Mental health | - |
dc.title | Modern Optimization for Statistics and Learning | - |
dc.type | Academic dissertations (Ph.D.) | - |
Appears in Collections: | Operations Research and Financial Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Eisenach_princeton_0181D_12945.pdf | 3.13 MB | Adobe PDF | View/Download |
Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.