αILP: thinking visual scenes as differentiable logic programs

Published: 01 Jan 2023, Last Modified: 15 Apr 2025Mach. Learn. 2023EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Deep neural learning has shown remarkable performance at learning representations for visual object categorization. However, deep neural networks such as CNNs do not explicitly encode objects and relations among them. This limits their success on tasks that require a deep logical understanding of visual scenes, such as Kandinsky patterns and Bongard problems. To overcome these limitations, we introduce \(\alpha {\textit{ILP}}\), a novel differentiable inductive logic programming framework that learns to represent scenes as logic programs—intuitively, logical atoms correspond to objects, attributes, and relations, and clauses encode high-level scene information. \(\alpha\)ILP has an end-to-end reasoning architecture from visual inputs. Using it, \(\alpha\)ILP performs differentiable inductive logic programming on complex visual scenes, i.e., the logical rules are learned by gradient descent. Our extensive experiments on Kandinsky patterns and CLEVR-Hans benchmarks demonstrate the accuracy and efficiency of \(\alpha {\textit{ILP}}\) in learning complex visual-logical concepts.
Loading