Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks. Lake, B. & Baroni, M. In Proceedings of the 35th International Conference on Machine Learning, pages 2873–2882, July, 2018. PMLR.
Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks [link]Paper  abstract   bibtex   8 downloads  
Humans can understand and produce new utterances effortlessly, thanks to their compositional skills. Once a person learns the meaning of a new verb "dax," he or she can immediately understand the meaning of "dax twice" or "sing and dax." In this paper, we introduce the SCAN domain, consisting of a set of simple compositional navigation commands paired with the corresponding action sequences. We then test the zero-shot generalization capabilities of a variety of recurrent neural networks (RNNs) trained on SCAN with sequence-to-sequence methods. We find that RNNs can make successful zero-shot generalizations when the differences between training and test commands are small, so that they can apply "mix-and-match" strategies to solve the task. However, when generalization requires systematic compositional skills (as in the "dax" example above), RNNs fail spectacularly. We conclude with a proof-of-concept experiment in neural machine translation, suggesting that lack of systematicity might be partially responsible for neural networks' notorious training data thirst.
@inproceedings{lakeGeneralizationSystematicityCompositional2018,
  title = {Generalization without {{Systematicity}}: {{On}} the {{Compositional Skills}} of {{Sequence-to-Sequence Recurrent Networks}}},
  shorttitle = {Generalization without {{Systematicity}}},
  booktitle = {Proceedings of the 35th {{International Conference}} on {{Machine Learning}}},
  author = {Lake, Brenden and Baroni, Marco},
  year = {2018},
  month = jul,
  pages = {2873--2882},
  publisher = {PMLR},
  issn = {2640-3498},
  url = {https://proceedings.mlr.press/v80/lake18a.html},
  urldate = {2024-03-18},
  abstract = {Humans can understand and produce new utterances effortlessly, thanks to their compositional skills. Once a person learns the meaning of a new verb "dax," he or she can immediately understand the meaning of "dax twice" or "sing and dax." In this paper, we introduce the SCAN domain, consisting of a set of simple compositional navigation commands paired with the corresponding action sequences. We then test the zero-shot generalization capabilities of a variety of recurrent neural networks (RNNs) trained on SCAN with sequence-to-sequence methods. We find that RNNs can make successful zero-shot generalizations when the differences between training and test commands are small, so that they can apply "mix-and-match" strategies to solve the task. However, when generalization requires systematic compositional skills (as in the "dax" example above), RNNs fail spectacularly. We conclude with a proof-of-concept experiment in neural machine translation, suggesting that lack of systematicity might be partially responsible for neural networks' notorious training data thirst.},
  langid = {english},
  file = {/Users/shanest/sync/library/Lake_Baroni/2018/Lake and Baroni - 2018 - Generalization without Systematicity On the Compo2.pdf;/Users/shanest/sync/library/Lake_Baroni/2018/Lake and Baroni - 2018 - Generalization without Systematicity On the Compo3.pdf}
}

Downloads: 8