Skip to content

Latest commit

 

History

History
40 lines (24 loc) · 3.8 KB

README.md

File metadata and controls

40 lines (24 loc) · 3.8 KB

MothMotifs

This project builds off of previous work published in [1] and uses pruned neural networks trained to model insect flight. A description of [1], as well as all code used to generate sparse neural networks for use in this project, can be found in this Github repository.

Background

Complex networks across many domains (ecological food webs, neural systems, electrical circuits) are made up of statistically significant, subgraphs called network motifs. Network motifs are described by the number of nodes they involve and the nature of the connections in-between the nodes (e.g., directed, or bi-directed). The order of the motif is defined by the number of nodes it involves (i.e. n-order motif involves n+1 nodes). In a feed-forward neural network, only directed n-order motifs are possible. Here are some of the 1st-, 2nd-, and 3rd-order motifs possible in a feed-forward neural network:

A subgraph must be statistically significant over a randomly connected graph to be considered a network motif of a given network. One metric for determining a subgraph’s statistical significance is its z-score when compared to randomly connected graphs.

Network motifs have been shown to be indicative of network functionality in control systems [2]. Quantifying the distribution of the network motifs of a sparse neural network trained for a control task (i.e., insect flight) may give us insight to the network substructures necessary for biological control tasks.

Project Description

Here, we are quantifying the network motif distribution over the sparse networks pruned in [1]. These networks were trained to model insect flight control and were sparsified via neural network pruning. We have developed our own subgraph counting algorithm based around using the masking matrices of the pruned networks. Network motifs are determined by calculating the z-score against random networks with the same number of nodes, connections, and layer structure. We use the motif z-score defined in [3]:

$$ \textrm{z-score} = \frac{N_{real} - N_{rand}}{\sigma} $$

where $N_{real}$ is the number of occurances of a subgraph in the real network (in this case a pruned NN), $N_{rand}$ is the average occurance of a subgraph in 1000 randomly generated networks, and $\sigma$ is the standard deviatation. The randomly generated networks have the same number of layers, nodes, and weights as the real pruned network. They also share all of the normal constraints of a feed-forward neural network (layer structure, bias term, etc.).

We are comparing the prevelence of network motifs across 400 networks trained to control insect flight. These networks were pruned to a maximum level of sparsity via a sequential, magnitude-based pruning algorithm described in [1].

References

[1] Zahn, Olivia, Jorge Bustamante Jr, Callin Switzer, Thomas L. Daniel, and J. Nathan Kutz. “Pruning deep neural networks generates a sparse, bio-inspired nonlinear controller for insect flight.” PLoS Computational Biology 18.9 (2022): e1010512. https://doi.org/10.1371/journal.pcbi.1010512

[2] Hu, Yu, et al. "Feedback through graph motifs relates structure and function in complex networks." Physical Review E 98.6 (2018): 062312. https://journals.aps.org/pre/pdf/10.1103/PhysRevE.98.062312

[3] Milo, Ron, et al. "Network motifs: simple building blocks of complex networks." Science 298.5594 (2002): 824-827. https://www.science.org/doi/pdf/10.1126/science.298.5594.824?casa_token=-0bkijkscyUAAAAA:teWQJRVXPBEQm29Ixu_zDD5stO6izuTdxi1rMzazg8nnnyJQtgl6Za-tTylPG6CLqsbJek0eXN8SK1c