Intuitively, compositional generalization is about combining things you know in new ways to solve a task, with little or no additional training. People can do this very well, even young children, but neural networks struggle. This is important because the cost to generalize without such a mechanism can be exponential in the number of variables in the task representation. Recent work evaluating neural networks for compositional generalization has mostly focused on natural language translation. In this presentation I'll review some of that work and talk about our experiments on composition in a completely geometric domain with no language but rather concepts specified in a first-order logical language which has richer constraints than those imposed by context free grammars. Our preliminary results indicate that neural nets do not generalize compositionally in this setting either.
Tim Klinger (IBM Research AI)
More from the Same Authors
2018 : Spotlights 2 »
Mausam · Ankit Anand · Parag Singla · Tarik Koc · Tim Klinger · Habibeh Naderi · Sungwon Lyu · Saeed Amizadeh · Kshitij Dwivedi · Songpeng Zu · Wei Feng · Balaraman Ravindran · Edouard Pineau · Abdulkadir Celikkanat · Deepak Venugopal
2017 : Poster session (and Coffee Break) »
Jacob Andreas · Kun Li · Conner Vercellino · Thomas Miconi · Wenpeng Zhang · Luca Franceschi · Zheng Xiong · Karim Ahmed · Laurent Itti · Tim Klinger · Mostafa Rohaninejad