Science of Computation and Language

Exhibition Program 7

Interpreting deep learning from network structure

Detecting communities in trained layered neural networks

Abstract

The effectiveness of layered neural networks is widely acknowledged for a wide range of tasks, including image recognition and natural language processing. However, the internal inference mechanism of layered neural networks is black-boxed, and we cannot understand the way in which they represent complex input-output relationships hidden in practical high-dimensional data sets. In this study, we propose a method for extracting the simplified global structure of a trained neural network by employing a network analysis method. Our proposed method first decomposes the units in each layer into communities according to their connection patterns with the adjacent layers, and then unites the multiple connections between a pair of communities into a single bundled connection. We aim for a future in which neural networks can be used over a wider application area than at present, such as automatic driving and medical care, where an explanation of the inference result is strongly required.

Reference

  • [1] C. Watanabe, K. Hiramatsu, K. Kashino, “Modular representation of layered neural networks,” Neural Networks, Vol. 97, pp. 62-73, 2018.
    [2] C. Watanabe, K. Hiramatsu, K. Kashino, “Community detection in layered neural networks based on signs of connection weights,” In Forum on Information Technology, 2017.
    [3] C. Watanabe, K. Hiramatsu, K. Kashino, “Recursive extraction of modular structure from layered neural networks using variational Bayes method,” in Proc. Discovery Science, Lecture Notes in Computer Science, Vol. 10558, pp. 207-222, 2017.

Poster

Photos

Presenters

Chihiro Watanabe)
Chihiro Watanabe
Media Information Laboratory