LEARNING NEUROSYMBOLIC GENERATIVE MODELS VIA PROGRAM SYNTHESISDownload PDF

Published: 10 Apr 2019, Last Modified: 05 May 2023drlStructPred 2019Readers: Everyone
Keywords: structure, deep learning, generative models, structured prediction
TL;DR: Applying program synthesis to the tasks of image completion and generation within a deep learning framework
Abstract: Significant strides have been made toward designing better generative models in recent years. Despite this progress, however, state-of-the-art approaches are still largely unable to capture complex global structure in data. For example, images of buildings typically contain spatial patterns such as windows repeating at regular intervals; state-of-the-art generative methods can’t easily reproduce these structures. We propose to address this problem by incorporating programs representing global structure into the generative model—e.g., a 2D for-loop may represent a configuration of windows. Furthermore, we propose a framework for learning these models by leveraging program synthesis to generate training data. On both synthetic and real-world data, we demonstrate that our approach is substantially better than the state-of-the-art at both generating and completing images that contain global structure.
5 Replies

Loading