Deep RNNs Encode Soft Hierarchical Syntax. Blevins, T., Levy, O., & Zettlemoyer, L. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), volume 2, pages 14–19, Stroudsburg, PA, USA, 2018. Association for Computational Linguistics. Paper doi abstract bibtex We present a set of experiments to demonstrate that deep recurrent neural networks (RNNs) learn internal representations that capture soft hierarchical notions of syntax from highly varied supervision. We consider four syntax tasks at different depths of the parse tree; for each word, we predict its part of speech as well as the first (parent), second (grandparent) and third level (great-grandparent) constituent labels that appear above it. These predictions are made from representations produced at different depths in networks that are pretrained with one of four objectives: dependency parsing, semantic role labeling, machine translation, or language modeling. In every case, we find a correspondence between network depth and syntactic depth, suggesting that a soft syntactic hierarchy emerges. This effect is robust across all conditions, indicating that the models encode significant amounts of syntax even in the absence of an explicit syntactic training supervision.
@inproceedings{Blevins2018,
abstract = {We present a set of experiments to demonstrate that deep recurrent neural networks (RNNs) learn internal representations that capture soft hierarchical notions of syntax from highly varied supervision. We consider four syntax tasks at different depths of the parse tree; for each word, we predict its part of speech as well as the first (parent), second (grandparent) and third level (great-grandparent) constituent labels that appear above it. These predictions are made from representations produced at different depths in networks that are pretrained with one of four objectives: dependency parsing, semantic role labeling, machine translation, or language modeling. In every case, we find a correspondence between network depth and syntactic depth, suggesting that a soft syntactic hierarchy emerges. This effect is robust across all conditions, indicating that the models encode significant amounts of syntax even in the absence of an explicit syntactic training supervision.},
address = {Stroudsburg, PA, USA},
archivePrefix = {arXiv},
arxivId = {1805.04218},
author = {Blevins, Terra and Levy, Omer and Zettlemoyer, Luke},
booktitle = {Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},
doi = {10.18653/v1/P18-2003},
eprint = {1805.04218},
file = {:Users/shanest/Documents/Library/Blevins, Levy, Zettlemoyer/Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2 Short Papers)/Blevins, Levy, Zettlemoyer - 2018 - Deep RNNs Encode Soft Hierarchical Syntax.pdf:pdf},
isbn = {9781948087346},
keywords = {method: diagnostic classifier,method: pre-training task comparison,phenomenon: hierarchical syntax},
pages = {14--19},
publisher = {Association for Computational Linguistics},
title = {{Deep RNNs Encode Soft Hierarchical Syntax}},
url = {http://aclweb.org/anthology/P18-2003},
volume = {2},
year = {2018}
}
Downloads: 0
{"_id":"ddXkPGyCJPPwto5JF","bibbaseid":"blevins-levy-zettlemoyer-deeprnnsencodesofthierarchicalsyntax-2018","authorIDs":[],"author_short":["Blevins, T.","Levy, O.","Zettlemoyer, L."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","abstract":"We present a set of experiments to demonstrate that deep recurrent neural networks (RNNs) learn internal representations that capture soft hierarchical notions of syntax from highly varied supervision. We consider four syntax tasks at different depths of the parse tree; for each word, we predict its part of speech as well as the first (parent), second (grandparent) and third level (great-grandparent) constituent labels that appear above it. These predictions are made from representations produced at different depths in networks that are pretrained with one of four objectives: dependency parsing, semantic role labeling, machine translation, or language modeling. In every case, we find a correspondence between network depth and syntactic depth, suggesting that a soft syntactic hierarchy emerges. This effect is robust across all conditions, indicating that the models encode significant amounts of syntax even in the absence of an explicit syntactic training supervision.","address":"Stroudsburg, PA, USA","archiveprefix":"arXiv","arxivid":"1805.04218","author":[{"propositions":[],"lastnames":["Blevins"],"firstnames":["Terra"],"suffixes":[]},{"propositions":[],"lastnames":["Levy"],"firstnames":["Omer"],"suffixes":[]},{"propositions":[],"lastnames":["Zettlemoyer"],"firstnames":["Luke"],"suffixes":[]}],"booktitle":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)","doi":"10.18653/v1/P18-2003","eprint":"1805.04218","file":":Users/shanest/Documents/Library/Blevins, Levy, Zettlemoyer/Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2 Short Papers)/Blevins, Levy, Zettlemoyer - 2018 - Deep RNNs Encode Soft Hierarchical Syntax.pdf:pdf","isbn":"9781948087346","keywords":"method: diagnostic classifier,method: pre-training task comparison,phenomenon: hierarchical syntax","pages":"14–19","publisher":"Association for Computational Linguistics","title":"Deep RNNs Encode Soft Hierarchical Syntax","url":"http://aclweb.org/anthology/P18-2003","volume":"2","year":"2018","bibtex":"@inproceedings{Blevins2018,\nabstract = {We present a set of experiments to demonstrate that deep recurrent neural networks (RNNs) learn internal representations that capture soft hierarchical notions of syntax from highly varied supervision. We consider four syntax tasks at different depths of the parse tree; for each word, we predict its part of speech as well as the first (parent), second (grandparent) and third level (great-grandparent) constituent labels that appear above it. These predictions are made from representations produced at different depths in networks that are pretrained with one of four objectives: dependency parsing, semantic role labeling, machine translation, or language modeling. In every case, we find a correspondence between network depth and syntactic depth, suggesting that a soft syntactic hierarchy emerges. This effect is robust across all conditions, indicating that the models encode significant amounts of syntax even in the absence of an explicit syntactic training supervision.},\naddress = {Stroudsburg, PA, USA},\narchivePrefix = {arXiv},\narxivId = {1805.04218},\nauthor = {Blevins, Terra and Levy, Omer and Zettlemoyer, Luke},\nbooktitle = {Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},\ndoi = {10.18653/v1/P18-2003},\neprint = {1805.04218},\nfile = {:Users/shanest/Documents/Library/Blevins, Levy, Zettlemoyer/Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2 Short Papers)/Blevins, Levy, Zettlemoyer - 2018 - Deep RNNs Encode Soft Hierarchical Syntax.pdf:pdf},\nisbn = {9781948087346},\nkeywords = {method: diagnostic classifier,method: pre-training task comparison,phenomenon: hierarchical syntax},\npages = {14--19},\npublisher = {Association for Computational Linguistics},\ntitle = {{Deep RNNs Encode Soft Hierarchical Syntax}},\nurl = {http://aclweb.org/anthology/P18-2003},\nvolume = {2},\nyear = {2018}\n}\n","author_short":["Blevins, T.","Levy, O.","Zettlemoyer, L."],"key":"Blevins2018","id":"Blevins2018","bibbaseid":"blevins-levy-zettlemoyer-deeprnnsencodesofthierarchicalsyntax-2018","role":"author","urls":{"Paper":"http://aclweb.org/anthology/P18-2003"},"keyword":["method: diagnostic classifier","method: pre-training task comparison","phenomenon: hierarchical syntax"],"metadata":{"authorlinks":{}},"downloads":0},"bibtype":"inproceedings","biburl":"https://www.shane.st/teaching/575/win20/MachineLearning-interpretability.bib","creationDate":"2020-01-09T19:12:32.385Z","downloads":0,"keywords":["method: diagnostic classifier","method: pre-training task comparison","phenomenon: hierarchical syntax"],"search_terms":["deep","rnns","encode","soft","hierarchical","syntax","blevins","levy","zettlemoyer"],"title":"Deep RNNs Encode Soft Hierarchical Syntax","year":2018,"dataSources":["okYcdTpf4JJ2zkj7A","znj7izS5PeehdLR3G"]}