Compositional Generalization in Semantic Parsing: Pre-training vs. Specialized Architectures

Daniel Peter Furrer
Nathan Scales
Nathanael Schärli
arXiv e-prints (2020), arXiv:2007.08970

Abstract

While mainstream machine learning methods are known to have limited ability to compositionally generalize, new architectures and techniques continue to be proposed to address this limitation. We investigate state-of-the-art techniques and architectures in order to assess their effectiveness in improving compositional generalization in semantic parsing tasks based on the SCAN and CFQ datasets.
We show that masked language model (MLM) pre-training rivals SCAN-inspired architectures on primitive holdout splits. On a more complex compositional task, we show that pre-training leads to significant improvements in performance vs. comparable non-pre-trained models, whereas architectures proposed to encourage compositional generalization on SCAN or in the area of algorithm learning fail to lead to significant improvements.
We establish a new state of the art on the CFQ compositional generalization benchmark using MLM pre-training together with an intermediate representation.