MARC보기
LDR00000nam u2200205 4500
001000000434482
00520200226151646
008200131s2019 ||||||||||||||||| ||eng d
020 ▼a 9781687975393
035 ▼a (MiAaPQ)AAI22620633
040 ▼a MiAaPQ ▼c MiAaPQ ▼d 247004
0820 ▼a 310
1001 ▼a Zhang, Ningshan.
24510 ▼a Essays in Applied Statistics and Machine Learning.
260 ▼a [S.l.]: ▼b New York University., ▼c 2019.
260 1 ▼a Ann Arbor: ▼b ProQuest Dissertations & Theses, ▼c 2019.
300 ▼a 184 p.
500 ▼a Source: Dissertations Abstracts International, Volume: 81-04, Section: B.
500 ▼a Advisor: Simonoff, Jeffrey S.
5021 ▼a Thesis (Ph.D.)--New York University, 2019.
506 ▼a This item must not be sold to any third party vendors.
520 ▼a In this dissertation, we look at three problems in applied statistics and machine learning. The first chapter considers the problem of fitting deeply nested hierarchical linear mixed models and its application to large scale recommender systems. We propose a recursive moment-based method for fitting hierarchical generalized linear mixed models of arbitrarily deep hierarchies. We show by simulations and a real world recommender system problem that, our proposed method is orders of magnitude faster than using off-the-shelf maximum likelihood procedures, while admitting comparable prediction performances.The second chapter examines the problem of joint modeling of longitudinal and time-to-event data via the latent class approach. Under the assumption that the longitudinal and time-to-event outcomes are independent conditioning on latent classes, we propose a nonparametric joint latent class modeling approach based on trees (JLCT). Simulation results as well as a real world example on the PAQUID dataset show that, the tree-based approach can be much faster than the most common parametric joint latent class modeling approach, the joint latent class model (JLCM). Furthermore, by using time-varying covariates in modeling survival risks and latent class memberships, JLCT can lead to a much more favorable prediction performance than JLCM, which is restricted to only using time-invariant covariates.In the last chapter, we study the multiple-source adaptation problem. We provide guarantees that there exists a distribution weighted combining rule that is robust with respect to any target mixture of the source distributions. These guarantees hold in the case where the conditional probabilities for the source domains are distinct, and where the loss function is the cross-entropy loss and the solution is normalized. Moreover, we give new algorithms for determining this robust combination solution for the cross-entropy loss and the squared loss. We report the results of a series of experiments with real-world datasets, where our algorithm outperforms competing approaches.
590 ▼a School code: 0146.
650 4 ▼a Statistics.
690 ▼a 0463
71020 ▼a New York University. ▼b Statistics.
7730 ▼t Dissertations Abstracts International ▼g 81-04B.
773 ▼t Dissertation Abstract International
790 ▼a 0146
791 ▼a Ph.D.
792 ▼a 2019
793 ▼a English
85640 ▼u http://www.riss.kr/pdu/ddodLink.do?id=T15493744 ▼n KERIS ▼z 이 자료의 원문은 한국교육학술정보원에서 제공합니다.
980 ▼a 202002 ▼f 2020
990 ▼a ***1008102
991 ▼a E-BOOK