Research

How Do Neural Sequence Models Generalize? Local and Global Cues for Out-of-Distribution Prediction

EMNLP

Authors

Published on

11/11/2021

After a neural sequence model encounters an unexpected token, can its behavior be predicted? We show that RNN and transformer language models exhibit structured, consistent generalization in out-of-distribution contexts. We begin by introducing two idealized models of generalization in next-word prediction: a lexical context model in which generalization is consistent with the last word observed, and a syntactic context model in which generalization is consistent with the global structure of the input. In experiments in English, Finnish, Mandarin, and random regular languages, we demonstrate that neural language models interpolate between these two forms of generalization: their predictions are well-approximated by a log-linear combination of lexical and syntactic predictive distributions. We then show that, in some languages, noise mediates the two forms of generalization: noise applied to input tokens encourages syntactic generalization, while noise in history representations encourages lexical generalization. Finally, we offer a preliminary theoretical explanation of these results by proving that the observed interpolation behavior is expected in log-linear models with a particular feature correlation structure. These results help explain the effectiveness of two popular regularization schemes and show that aspects of sequence model generalization can be understood and controlled.

Please cite our work using the BibTeX below.

@inproceedings{bau-andreas-2021-neural,
    title = "How Do Neural Sequence Models Generalize? Local and Global Cues for Out-of-Distribution Prediction",
    author = "Bau, D. Anthony  and
      Andreas, Jacob",
    booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2021",
    address = "Online and Punta Cana, Dominican Republic",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.emnlp-main.448",
    doi = "10.18653/v1/2021.emnlp-main.448",
    pages = "5513--5526",
}
Close Modal