MARC보기
LDR00000nam u2200205 4500
001000000435293
00520200228092520
008200131s2019 ||||||||||||||||| ||eng d
020 ▼a 9781085628105
035 ▼a (MiAaPQ)AAI13880243
040 ▼a MiAaPQ ▼c MiAaPQ ▼d 247004
0820 ▼a 001
1001 ▼a Eisenach, Carson Mark.
24510 ▼a Modern Optimization for Statistics and Learning.
260 ▼a [S.l.]: ▼b Princeton University., ▼c 2019.
260 1 ▼a Ann Arbor: ▼b ProQuest Dissertations & Theses, ▼c 2019.
300 ▼a 215 p.
500 ▼a Source: Dissertations Abstracts International, Volume: 81-02, Section: B.
500 ▼a Advisor: Liu, Han.
5021 ▼a Thesis (Ph.D.)--Princeton University, 2019.
506 ▼a This item must not be sold to any third party vendors.
520 ▼a Traditional problems in statistics and machine learning are relatively well understood - they often feature low dimensionality, convex loss functions, and independent, identically distributed data. By contrast, many modern learning problems feature high dimensional data, non-convex learning objectives, and data distributions that change during the learning process. Whether the problem of interest is labeled as statistics, machine learning, statistical learning, or reinforcement learning, methods for solving it can be viewed as the stochastic optimization of some objective function.Accordingly, we address the aforementioned challenges via the lens of statistical optimization - a statistical approach for understanding and solving stochastic optimization. In particular, we focus on deriving new methodology with computational and statistical guarantees for two classes of problems: recovering and performing inference on latent patterns in high-dimensional graphical models, and continuous control over bounded action spaces.In the first part of this dissertation, we consider a class of cluster-based graphical models. We introduce a novel algorithm for variable clustering named FORCE, based on solving a convex relaxation of the K-means criterion, as well as post-dimension reduction inferential procedures. In the second part, we consider the reinforcement learning (RL) setting, where an agent seeks to learn a decision-making policy based on feedback from its environment. We derive a novel class of variance-reduced estimators called Marginal Policy Gradients, and demonstrate both their improved statistical properties and their application to several control tasks.
590 ▼a School code: 0181.
650 4 ▼a Statistics.
650 4 ▼a Computer science.
650 4 ▼a Artificial intelligence.
690 ▼a 0463
690 ▼a 0984
690 ▼a 0800
71020 ▼a Princeton University. ▼b Operations Research and Financial Engineering.
7730 ▼t Dissertations Abstracts International ▼g 81-02B.
773 ▼t Dissertation Abstract International
790 ▼a 0181
791 ▼a Ph.D.
792 ▼a 2019
793 ▼a English
85640 ▼u http://www.riss.kr/pdu/ddodLink.do?id=T15491156 ▼n KERIS ▼z 이 자료의 원문은 한국교육학술정보원에서 제공합니다.
980 ▼a 202002 ▼f 2020
990 ▼a ***1816162
991 ▼a E-BOOK