[1] LECUN Y A,BOTTOU L,ORR G B,et al. Efficient backprop[M]//ORR G B,MÜLLER K R. Neural Networks:Tricks of the Trade,LNCS 7700. Berlin:Springer,2012:9-48. [2] HASSIBI B,STORK D G. Second order derivatives for network pruning:optimal brain surgeon[C]//Proceedings of the 5th International Conference on Neural Information Processing Systems. San Francisco:Morgan Kaufmann Publishers Inc.,1992:164-171. [3] SRINIVAS S,BADU R V. Learning neural network architectures using backpropagation[C]//Proceedings of the 2016 British Machine Vision Conference. Durham:BMVA Press,2016:No. 104. [4] WEN W,WU C,WANG Y,et al. Learning structured sparsity in deep neural networks[C]//Proceedings of the 30th International Conference on Neural Information Processing System. New York:Curran Associates Inc.,2016:2082-2090. [5] LI H,KADAV A,DURDANOVIC I,et al. Pruning filters for efficient convnets[EB/OL].[2019-06-20]. https://arxiv.org/pdf/1608.08710.pdf. [6] GUPTA S,AGRAWAL A,GOPALAKRISHNAN K,et al. Deep learning with limited numerical precision[C]//Proceedings of the 32nd International Conference on Machine Learning. New York:JMLR.org,2015:1737-1746. [7] GYSEL P,MOTAMEDI M,GHIASI S. Hardware-oriented approximation of convolutional neural networks[EB/OL].[2019-06-20]. https://arxiv.org/pdf/1604.03168.pdf. [8] DENIL M,SHAKIBI B,DINH L,et al. Predicting parameters in deep learning[C]//Proceedings of the 26th International Conference on Neural Information Processing Systems. New York:Curran Associates Inc.,2013:2148-2516. [9] JADERBERG M,VEDALI A,ZISSERMAN A. Speeding up convolutional neural networks with low rank expansions[C]//Proceedings of the 2014 British Machine Vision Conference. Durham:BMVA Press,2014:No. 73. [10] DENTON E,ZAZREMBA W,BRUNA J,et al. Exploiting linear structure within convolutional networks for efficient evaluation[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems. Cambridge:MIT Press, 2014:1269-1277. [11] BUCILUǍ C,CARUANA R,NICULESCU-MIZIL A. Model compression[C]//Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York:ACM,2006:535-541. [12] HINTON G,VINYALS O,DEAN J. Distilling the knowledge in a neural network[EB/OL].[2019-06-20]. https://arxiv.org/pdf/1503.02531.pdf. [13] HUANG Z,WANG N. Like what you like:knowledge distill via neuron selectivity transfer[EB/OL].[2019-06-20]. https://arxiv.org/pdf/1707.01219.pdf. [14] LI D,WANG X,KONG D. DeepRebirth:accelerating deep neural network execution on mobile devices[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence. Palo Alto,CA:AAAI Press,2018:2322-2330. [15] LECUN Y,BOTTOU L,BENGIO Y,et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998,86(11):2278-2324. [16] 黄文坚, 唐源. TesnsorFlow实战[M]. 北京:电子工业出版社, 2017:233-242. (HUANG W J,TANG Y. Actual Combat of TensorFlow[M]. Beijing:Publishing House of Electronics Industry, 2017:233-242.) [17] HAN S,POOL J,TRAN J,et al. Learning both weights and connections for efficient neural networks[C]//Proceedings of the 28th International Conference on Neural Information Processing Systems. Cambridge:MIT Press,2015:1135-1143. [18] LIU Z,LI J,SHEN Z,et al. Learning efficient convolutional networks through network slimming[C]//Proceedings of the 2017 International Conference on Computer Vision. Piscataway:IEEE, 2017:2755-2763. [19] 靳丽蕾, 杨文柱, 王思乐, 等. 一种用于卷积神经网络压缩的混合剪枝方法[J]. 小型微型计算机系统,2018,39(12):2596-2601. (JIN L L,YANG W Z,WANG S L. et al. Mixed pruning method for convolution neural network compression[J]. Journal of Chinese Computer Systems,2018,39(12):2596-2601. [20] FRANKLE J,CARBIN M. The lottery ticket hypothesis:finding sparse trainable neural networks[EB/OL].[2019-06-20]. https://arxiv.org/pdf/1803.03635.pdf. |