MARC보기
LDR00000nam u2200205 4500
001000000435646
00520200228102107
008200131s2019 ||||||||||||||||| ||eng d
020 ▼a 9781392688106
035 ▼a (MiAaPQ)AAI27539403
040 ▼a MiAaPQ ▼c MiAaPQ ▼d 247004
0820 ▼a 001
1001 ▼a Gonzalez Martinez, Pablo .
24510 ▼a Do It Like a Syntactician: Using Binary Grammaticality Judgements to Train Sentence Encoders and Assess Their Sensitivity to Syntactic Structure.
260 ▼a [S.l.]: ▼b City University of New York., ▼c 2019.
260 1 ▼a Ann Arbor: ▼b ProQuest Dissertations & Theses, ▼c 2019.
300 ▼a 105 p.
500 ▼a Source: Dissertations Abstracts International, Volume: 81-04, Section: B.
500 ▼a Advisor: Sakas, William.
5021 ▼a Thesis (Ph.D.)--City University of New York, 2019.
506 ▼a This item must not be sold to any third party vendors.
520 ▼a The binary nature of grammaticality judgments and their use to access the structure of syntax are a staple of modern linguistics. However, computational models of natural language rarely make use of grammaticality in their training or application. Furthermore, developments in modern neural NLP have produced a myriad of methods that push the baselines in many complex tasks, but those methods are typically not evaluated from a linguistic perspective. In this dissertation I use grammaticality judgements with artificially generated ungrammatical sentences to assess the performance of several neural encoders and propose them as a suitable training target to make models learn specific syntactic rules. I generate artificial ungrammatical sentences via two methods. First by randomly pulling words following the n-gram distribution of a corpus of real sentences (I call these Word salads). Second, by corrupting sentences from a real corpus by altering them (changing verbal or adjectival agreement or removing the main verb). We then train models with an encoder using word embeddings and long short term memory (LSTMs) to discriminate between real sentences and ungrammatical sentences. We show that word salads can be distinguished by the model well for low order n-grams but that the model does not generalize well for higher orders. Furthermore, the word salads do not help the model in recognizing corrupted sentences. We then test the contributions of pre-trained word embeddings, deep LSTM and bidirectional LSTM. We find that the biggest contribution is adding pre-trained word embeddings. We also find that additional layers contribute differently to the performance of unidirectional and bidirectional models and that deeper models have more performance variability across training runs.
590 ▼a School code: 0046.
650 4 ▼a Linguistics.
650 4 ▼a Computer science.
650 4 ▼a Artificial intelligence.
690 ▼a 0290
690 ▼a 0984
690 ▼a 0800
71020 ▼a City University of New York. ▼b Linguistics.
7730 ▼t Dissertations Abstracts International ▼g 81-04B.
773 ▼t Dissertation Abstract International
790 ▼a 0046
791 ▼a Ph.D.
792 ▼a 2019
793 ▼a English
85640 ▼u http://www.riss.kr/pdu/ddodLink.do?id=T15494360 ▼n KERIS ▼z 이 자료의 원문은 한국교육학술정보원에서 제공합니다.
980 ▼a 202002 ▼f 2020
990 ▼a ***1816162
991 ▼a E-BOOK