Timezone: »

Formalizing Consistency and Coherence of Representation Learning
Harald Strömfelt · Luke Dickens · Artur Garcez · Alessandra Russo

Tue Nov 29 02:00 PM -- 04:00 PM (PST) @ Hall J #916

In the study of reasoning in neural networks, recent efforts have sought to improve consistency and coherence of sequence models, leading to important developments in the area of neuro-symbolic AI. In symbolic AI, the concepts of consistency and coherence can be defined and verified formally, but for neural networks these definitions are lacking. The provision of such formal definitions is crucial to offer a common basis for the quantitative evaluation and systematic comparison of connectionist, neuro-symbolic and transfer learning approaches. In this paper, we introduce formal definitions of consistency and coherence for neural systems. To illustrate the usefulness of our definitions, we propose a new dynamic relation-decoder model built around the principles of consistency and coherence. We compare our results with several existing relation-decoders using a partial transfer learning task based on a novel data set introduced in this paper. Our experiments show that relation-decoders that maintain consistency over unobserved regions of representation space retaincoherence across domains, whilst achieving better transfer learning performance.

Author Information

Harald Strömfelt (Imperial College London)
Luke Dickens (University College London)
Artur Garcez (City, University of London)
Alessandra Russo (Imperial College London)

More from the Same Authors