A mechanistically interpretable neural-network architecture for discovery of regulatory genomics

Published: 04 Mar 2024, Last Modified: 24 Apr 2024MLGenX 2024 SpotlightEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Mechanistic interpretability, regulatory genomics, motif discovery, explainability
TL;DR: We designed a mechanistically interpretable neural network which reveals regulatory motifs and syntax directly in its weights and activations
Abstract: Deep neural networks have shown unparalleled success in mapping genomic DNA sequences to associated readouts such as protein–DNA binding. Beyond prediction, the goal of these networks is to then learn the underlying motifs (and their syntax) which drive genome regulation. Traditionally, this has been done by applying fragile and computationally expensive post-hoc analysis pipelines to trained models. Instead, we propose an entirely alternative method for learning motif biology from neural networks. We designed a mechanistically interpretable neural-network architecture for regulatory genomics, where motifs and their syntax are directly encoded and readable from the learned weights and activations, thus eliminating the need for post-hoc pipelines. Our model is also more robust to variable sequence contexts and against adversarial attacks, while attaining predictive performance comparable to its traditional counterparts.
Submission Number: 42
Loading