Improving Convolutional Network Interpretability with Exponential Activations

Koo, Peter K., Ploenzke, Matt (2019) Improving Convolutional Network Interpretability with Exponential Activations. bioRxiv. p. 650804. (Unpublished)

DOI: 10.1101/650804


Deep convolutional networks trained on regulatory genomic sequences tend to learn distributed representations of sequence motifs across many first layer filters. This makes it challenging to decipher which features are biologically meaningful. Here we introduce the exponential activation that – when applied to first layer filters – leads to more interpretable representations of motifs, both visually and quantitatively, compared to rectified linear units. We demonstrate this on synthetic DNA sequences which have ground truth with various convolutional networks, and then show that this phenomenon holds on in vivo DNA sequences.

Item Type: Paper
Subjects: bioinformatics > computational biology
organs, tissues, organelles, cell types and functions > tissues types and functions > neural networks
CSHL Authors:
Communities: CSHL labs > Koo Lab
Depositing User: Matthew Dunn
Date: 2019
Date Deposited: 17 Sep 2019 15:07
Last Modified: 17 Sep 2019 15:07

Actions (login required)

Administrator's edit/view item Administrator's edit/view item
CSHL HomeAbout CSHLResearchEducationNews & FeaturesCampus & Public EventsCareersGiving