LEARNING NEUROSYMBOLIC GENERATIVE MODELS VIA PROGRAM SYNTHESIS

Halley Young, Osbert Bastani, Mayur Naik

Mar 23, 2019 ICLR 2019 Workshop drlStructPred Blind Submission readers: everyone
  • Keywords: structure, deep learning, generative models, structured prediction
  • TL;DR: Applying program synthesis to the tasks of image completion and generation within a deep learning framework
  • Abstract: Significant strides have been made toward designing better generative models in recent years. Despite this progress, however, state-of-the-art approaches are still largely unable to capture complex global structure in data. For example, images of buildings typically contain spatial patterns such as windows repeating at regular intervals; state-of-the-art generative methods can’t easily reproduce these structures. We propose to address this problem by incorporating programs representing global structure into the generative model—e.g., a 2D for-loop may represent a configuration of windows. Furthermore, we propose a framework for learning these models by leveraging program synthesis to generate training data. On both synthetic and real-world data, we demonstrate that our approach is substantially better than the state-of-the-art at both generating and completing images that contain global structure.
0 Replies

Loading