Neural language models as psycholinguistic subjects: Representations of syntactic state. Futrell, R., Wilcox, E., Morita, T., Qian, P., Ballesteros, M., & Levy, R. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 32–42, Stroudsburg, PA, USA, 2019. Association for Computational Linguistics.
Neural language models as psycholinguistic subjects: Representations of syntactic state [link]Paper  doi  abstract   bibtex   
We deploy the methods of controlled psycholinguistic experimentation to shed light on the extent to which the behavior of neural network language models reflects incremental representations of syntactic state. To do so, we examine model behavior on artificial sentences containing a variety of syntactically complex structures. We test four models: two publicly available LSTM sequence models of English (Jozefowicz et al., 2016; Gulordava et al., 2018) trained on large datasets; an RNNG (Dyer et al., 2016) trained on a small, parsed dataset; and an LSTM trained on the same small corpus as the RNNG. We find evidence that the LSTMs trained on large datasets represent syntactic state over large spans of text in a way that is comparable to the RNNG, while the LSTM trained on the small dataset does not or does so only weakly.
@inproceedings{Futrell2019,
abstract = {We deploy the methods of controlled psycholinguistic experimentation to shed light on the extent to which the behavior of neural network language models reflects incremental representations of syntactic state. To do so, we examine model behavior on artificial sentences containing a variety of syntactically complex structures. We test four models: two publicly available LSTM sequence models of English (Jozefowicz et al., 2016; Gulordava et al., 2018) trained on large datasets; an RNNG (Dyer et al., 2016) trained on a small, parsed dataset; and an LSTM trained on the same small corpus as the RNNG. We find evidence that the LSTMs trained on large datasets represent syntactic state over large spans of text in a way that is comparable to the RNNG, while the LSTM trained on the small dataset does not or does so only weakly.},
address = {Stroudsburg, PA, USA},
archivePrefix = {arXiv},
arxivId = {1903.03260},
author = {Futrell, Richard and Wilcox, Ethan and Morita, Takashi and Qian, Peng and Ballesteros, Miguel and Levy, Roger},
booktitle = {Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)},
doi = {10.18653/v1/N19-1004},
eprint = {1903.03260},
file = {:Users/shanest/Documents/Library/Futrell et al/Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics Human Language Technologie./Futrell et al. - 2019 - Neural language models as psycholinguistic subjects Representations of syntactic state.pdf:pdf},
keywords = {method: psycholinguistic,phenomenon: incremental syntax},
pages = {32--42},
publisher = {Association for Computational Linguistics},
title = {{Neural language models as psycholinguistic subjects: Representations of syntactic state}},
url = {http://aclweb.org/anthology/N19-1004},
year = {2019}
}

Downloads: 0