AUTHOR=Zhao Feifei , Zeng Yi TITLE=Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain JOURNAL=Frontiers in Systems Neuroscience VOLUME=15 YEAR=2021 URL=https://www.frontiersin.org/journals/systems-neuroscience/articles/10.3389/fnsys.2021.620558 DOI=10.3389/fnsys.2021.620558 ISSN=1662-5137 ABSTRACT=
Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to be deployed on mobile devices. Dynamically optimizing the network architecture by pruning unused synapses is a promising technique for solving this problem. Most existing pruning methods focus on reducing the redundancy of deep convolutional neural networks by pruning unimportant filters or weights, at the cost of accuracy drop. In this paper, we propose an effective brain-inspired synaptic pruning method to dynamically modulate the network architecture and simultaneously improve network performance. The proposed model is biologically inspired as it dynamically eliminates redundant connections based on the synaptic pruning rules used during the brain's development. Connections are pruned if they are not activated or less activated multiple times consecutively. Extensive experiments demonstrate the effectiveness of our method on classification tasks of different complexity with the MNIST, Fashion MNIST, and CIFAR-10 datasets. Experimental results reveal that even for a compact network, the proposed method can also remove up to 59–90% of the connections, with relative improvement in learning speed and accuracy.