- Efficient processing of deep neural networks: A tutorial and survey. Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang, and Joel Emer. paper
- A survey of model compression and acceleration for deep neural networks. Yu Cheng, Duo Wang, Pan Zhou, and Tao Zhang. paper
- An analysis of deep neural network models for practical applications. Alfredo Canziani, Adam Paszke, and Eugenio Culurciello. paper
- Deep convolutional neural networks for image classiffication: A comprehensive review. Waseem Rawat and Zenghui Wang. paper
-
-
-
-
- Distilling the knowledge in a neural network. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. paper
- Fitnets: Hints for thin deep nets. Adriana Romero, Nicolas Ballas, Samira Ebrahimi Kahou, Antoine Chassang, Carlo Gatta, and Yoshua Bengio.
- Harnessing deep neural networks with logic rules. Zhiting Hu, Xuezhe Ma, Zhengzhong Liu, Eduard Hovy, and Eric Xing.
- Do deep nets really need to be deep? Jimmy Ba and Rich Caruana.
- Do deep convolutional nets really need to be deep and convolutional? Gregor Urban, Krzysztof J Geras, Samira Ebrahimi Kahou, Ozlem Aslan, Shengjie Wang, Rich Caruana, Abdelrahman Mohamed, Matthai Philipose, and Matt Richardson.
- Transferring knowledge from a rnn to a dnn. William Chan, Nan Rosemary Ke, and Ian Lane.
- Face model compression by distilling knowledge from neurons. Ping Luo, Zhenyao Zhu, Ziwei Liu, Xiaogang Wang, Xiaoou Tang, et al.
- Like what you like: Knowledge distill via neuron selectivity transfer. Zehao Huang and Naiyan Wang.
- Darkrank: Accelerating deep metric learning via cross sample similarities transfer. Yuntao Chen, Naiyan Wang, and Zhaoxiang Zhang.
- Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. Sergey Zagoruyko and Nikos Komodakis.
- Accelerating convolutional neural networks with dominant convolutional kernel and knowledge pre-regression. Zhenyang Wang, Zhidong Deng, and Shiyao Wang.
- Rocket launching: A universal and efficient framework for training well-performing light net. Guorui Zhou, Ying Fan, Runpeng Cui, Weijie Bian, Xiaoqiang Zhu, and Kun Gai.
- Optimal brain damage. Yann LeCun, John S Denker, and Sara A Solla.
- Second order derivatives for network pruning: Optimal brain surgeon. Babak Hassibi, David G Stork, et al.
- Learning both weights and connections for efficient neural network. Song Han, Jeff Pool, John Tran, and William Dally.
- Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. Song Han, Huizi Mao, and William J Dally.
- Dsd: Dense-sparse-dense training for deep neural networks. Song Han, Jeff Pool, Sharan Narang, Huizi Mao, Enhao Gong, Shijian Tang, Erich Elsen, Peter Vajda, Manohar Paluri, John Tran, et al.
- Data-free parameter pruning for deep neural networks. Suraj Srinivas and R Venkatesh Babu.
- Dynamic network surgery for efficient dnns. Yiwen Guo, Anbang Yao, and Yurong Chen.
- Faster cnns with direct sparse convolutions and guided pruning. Jongsoo Park, Sheng Li, Wei Wen, Ping Tak Peter Tang, Hai Li, Yiran Chen, and Pradeep Dubey.
- Pruning convolutional neural networks for resource efficient transfer learning. Pavlo Molchanov, Stephen Tyree, Tero Karras, Timo Aila, and Jan Kautz.
- Pruning filters for efficient convnets. Hao Li, Asim Kadav, Igor Durdanovic, Hanan Samet, and Hans Peter Graf.
- Structured pruning of deep convolutional neural networks. Sajid Anwar, Kyuyeon Hwang, and Wonyong Sung.
- A simple yet effective method to prune dense layers of neural networks. Mohammad Babaeizadeh, Paris Smaragdis, and Roy H Campbell.
- Designing energy-efficient convolutional neural networks using energy-aware pruning. Tien-Ju Yang, Yu-Hsin Chen, and Vivienne Sze.
- Bayesian compression for deep learning. Christos Louizos, Karen Ullrich, and Max Welling.
- Weight uncertainty in neural networks. Charles Blundell, Julien Cornebise, Koray Kavukcuoglu, and Daan Wierstra.
- An entropy-based pruning method for cnn compression. Jian-Hao Luo and Jianxin Wu.
- Exploring the regularity of sparse structure in convolutional neural networks. Huizi Mao, Song Han, Jeff Pool, Wenshuo Li, Xingyu Liu, Yu Wang, and William J Dally.
- Fast convnets using group-wise brain damage. Vadim Lebedev and Victor Lempitsky.
- Thinet: A filter level pruning method for deep neural network compression. Jian-Hao Luo, Jianxin Wu, and Weiyao Lin.
- Network trimming: A data-driven neuron pruning approach towards efficient deep architectures. Hengyuan Hu, Rui Peng, Yu-Wing Tai, and Chi-Keung Tang.
- Accelerating deep learning with shrinkage and recall. Shuai Zheng, Abhinav Vishnu, and Chris Ding.
- Prune the convolutional neural networks with sparse shrink. Xin Li and Changsong Liu.
- Neuron pruning for compressing deep networks using maxout architectures. Fernando Moya Rueda, Rene Grzeszick, and Gernot A Fink.
- Fine-pruning: Joint fine-tuning and compression of a convolutional network with bayesian optimization. Frederick Tung, Srikanth Muralidharan, and Greg Mori.
- Structured bayesian pruning via log-normal multiplicative noise. Kirill Neklyudov, Dmitry Molchanov, Arsenii Ashukha, and Dmitry Vetrov.
- Towards evolutional compression. Yunhe Wang, Chang Xu, Jiayan Qiu, Chao Xu, and Dacheng Tao.
- Lazy evaluation of convolutional filters. Sam Leroux, Steven Bohez, Cedric De Boom, Elias De Coninck, Tim Verbelen, Bert Vankeirsbilck, Pieter Simoens, and Bart Dhoedt.
- Sparsely-connected neural networks: Towards efficient vlsi implementation of deep neural networks. Arash Ardakani, Carlo Condo, andWarren J Gross.
- Net-trim: A layer-wise convex pruning of deep neural networks. Alireza Aghasi, Nam Nguyen, and Justin Romberg.
- Learning with confident examples: Rank pruning for robust classification with noisy labels. Curtis G. Northcutt, Tailin Wu, and Isaac L. Chuang.
- Compact deep convolutional neural networks with coarse pruning. Sajid Anwar and Wonyong Sung.
- Towards thinner convolutional neural networks through gradually global pruning. Zhengtao Wang, Ce Zhu, Zhiqiang Xia, Qi Guo, and Yipeng Liu.
- The incredible shrinking neural network: New perspectives on learning representations through the lens of pruning. Nikolas Wolfe, Aditya Sharma, Lukas Drude, and Bhiksha Raj.
- Training skinny deep neural networks with iterative hard thresholding methods. Xiaojie Jin, Xiaotong Yuan, Jiashi Feng, and Shuicheng Yan.
- Reducing the model order of deep neural networks using information theory. Ming Tu, Visar Berisha, Yu Cao, and Jae Sun Seo.
- Learning efficient convolutional networks through network slimming. Zhuang Liu, Jianguo Li, Zhiqiang Shen, Gao Huang, Shoumeng Yan, and Changshui Zhang.
- Channel pruning for accelerating very deep neural networks. Yihui He, Xiangyu Zhang, and Jian Sun.
- Incomplete dot products for dynamic computation scaling in neural network inference. H. T. Kung Bradley McDanel, Surat Teerapittayanon.
- To prune, or not to prune: exploring the efficacy of pruning for model compression. Michael Zhu and Suyog Gupta.
- Data-driven sparse structure selection for deep neural networks. Zehao Huang and Naiyan Wang.
- Pruning convnets online for efficient specialist models. Jia Guo and Miodrag Potkonjak.
- Sparse convolutional neural networks. Baoyuan Liu, Min Wang, Hassan Foroosh, Marshall Tappen, and Marianna Pensky.
- Group sparse regularization for deep neural networks. Simone Scardapane, Danilo Comminiello, Amir Hussain, and Aurelio Uncini.
- The power of sparsity in convolutional neural networks. Soravit Changpinyo, Mark Sandler, and Andrey Zhmoginov.
- Spatially-sparse convolutional neural networks. Benjamin Graham.
- Shakeout: A new approach to regularized deep neural network training. Guoliang Kang, Jun Li, and Dacheng Tao.
- Sparse activity and sparse connectivity in supervised learning. Markus Thom and Gunther Palm.
- Learning structured sparsity in deep neural networks. Wei Wen, Chunpeng Wu, Yandan Wang, Yiran Chen, and Hai Li.
- Perforatedcnns: Acceleration through elimination of redundant convolutions. Mikhail Figurnov, Aizhan Ibraimova, Dmitry P Vetrov, and Pushmeet Kohli.
- Training compressed fully-connected networks with a density-diversity penalty. Shengjie Wang, Haoran Cai, Jeff Bilmes, and William Noble.
- Stochasticnet: Forming deep neural networks via stochastic connectivity. Mohammad Javad Shafiee, Parthipan Siva, and Alexander Wong.
- Deep roots: Improving cnn efficiency with hierarchical filter groups. Yani Ioannou, Duncan Robertson, Roberto Cipolla, and Antonio Criminisi.
- Less is more: Towards compact cnns. Hao Zhou, Jose M Alvarez, and Fatih Porikli.
- More is less: A more complicated network with less inference complexity. Xuanyi Dong, Junshi Huang, Yi Yang, and Shuicheng Yan.
- Memory bounded deep convolutional networks. Maxwell D Collins and Pushmeet Kohli.
- Combined group and exclusive sparsity for deep neural networks. Jaehong Yoon and Sung Ju Hwang.
- On compressing deep models by low rank and sparse decomposition. Xiyu Yu, Tongliang Liu, Xinchao Wang, and Dacheng Tao.
- Speeding up convolutional neural networks by exploiting the sparsity of rectifier units. Shaohuai Shi and Xiaowen Chu.
- Alternating direction method of multipliers for sparse convolutional neural networks. Farkhondeh Kiaee, Christian Gagn, and Mahdieh Abbasi.
- Training sparse neural networks. Suraj Srinivas, Akshayvarun Subramanya, and R. Venkatesh Babu.
- Dyvedeep: Dynamic variable effort deep neural networks. Balaraman Ravindran Anand Raghunathan SanjayGanapathy, Swagath Venkataramani.
- Freezeout: Accelerate training by progressively freezing layers. Andrew Brock, Theodore Lim, J. M. Ritchie, and Nick Weston.
- Convolutional neural networks at constrained time cost. Kaiming He and Jian Sun.
-
-
-
-
- Fixed-point feedforward deep neural network design using weights+1, 0, and- 1. Kyuyeon Hwang and Wonyong Sung.
- Fixed point quantization of deep convolutional networks. Darryl Lin, Sachin Talathi, and Sreekanth Annapureddy.
- Binaryconnect: Training deep neural networks with binary weights during propagations. Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David.
- Binarized neural networks: Training deep neural networks with weights and activations constrained to+ 1 or-1. Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio.
- Quantized neural networks: Training neural networks with low precision weights and activations. Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio.
- Xnor-net: Imagenet classification using binary convolutional neural networks. Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi.
- Bitwise neural networks. Minje Kim and Paris Smaragdis.
- Training quantized nets: A deeper understanding. Hao Li, Soham De, Zheng Xu, Christoph Studer, Hanan Samet, and Tom Goldstein.
- Shiftcnn: Generalized low-precision architecture for inference of convolutional neural networks. Denis A. Gudovskiy and Luca Rigazio.
- Gated xnor networks: Deep neural networks with ternary weights and activations under a unified discretization framework. Lei Deng, Peng Jiao, Jing Pei, Zhenzhi Wu, and Guoqi Li.
- The high-dimensional geometry of binary neural networks. Alexander G Anderson and Cory P Berg.
- Compressing deep convolutional networks using vector quantization. Yunchao Gong, Liu Liu, Ming Yang, and Lubomir Bourdev.
- Compression of deep neural networks on the fly. Guillaume Soulie, Vincent Gripon, and Maelys Robert.
- Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou.
- Ternary weight networks. Fengfu Li, Bo Zhang, and Bin Liu.
- Trained ternary quantization. Chenzhuo Zhu, Song Han, Huizi Mao, and William J Dally.
- Incremental network quantization: Towards lossless cnns with low-precision weights. Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen.
- Quantized convolutional neural networks for mobile devices. Jiaxiang Wu, Cong Leng, Yuhang Wang, Qinghao Hu, and Jian Cheng.
- Compressing neural networks with the hashing trick. Wenlin Chen, James Wilson, Stephen Tyree, Kilian Weinberger, and Yixin Chen.
- Scalable and sustainable deep learning via randomized hashing. Ryan Spring and Anshumali Shrivastava.
- Functional hashing for compressing neural networks. Lei Shi, Shikun Feng, et al.
- Compressing convolutional neural networks. Wenlin Chen, James T Wilson, Stephen Tyree, Kilian Q Weinberger, and Yixin Chen.
- Compressing convolutional neural networks in the frequency domain. Wenlin Chen, James Wilson, Stephen Tyree, Kilian Q. Weinberger, and Yixin Chen.
- Neural networks with few multiplications. Zhouhan Lin, Matthieu Courbariaux, Roland Memisevic, and Yoshua Bengio.
- Training binary multilayer neural networks for image classification using expectation backpropagation. Zhiyong Cheng, Daniel Soudry, Zexi Mao, and Zhenzhong Lan.
- Improving the speed of neural networks on cpus. Vincent Vanhoucke, Andrew Senior, and Mark Z Mao.
- Soft weight-sharing for neural network compression. Karen Ullrich, Edward Meeds, and Max Welling.
- Towards the limit of network quantization. Yoojin Choi, Mostafa El-Khamy, and Jungwon Lee.
- Tensorizing neural networks. Alexander Novikov, Dmitrii Podoprikhin, Anton Osokin, and Dmitry P Vetrov.
- Deep learning with limited numerical precision. Suyog Gupta, Ankur Agrawal, Kailash Gopalakrishnan, and Pritish Narayanan.
- Finite precision error analysis of neural network hardware implementations. Jordan L Holi and J-N Hwang.
- Deep learning with low precision by halfwave gaussian quantization. Zhaowei Cai, Xiaodong He, Jian Sun, and Nuno Vasconcelos.
- Deep quantization: Encoding convolutional activations with deep generative model. Zhaofan Qiu, Ting Yao, and Tao Mei.
- Weighted-entropy-based quantization for deep neural networks. Junwhan Ahn Eunhyeok Park and Sungjoo Yoo.
- Extremely low bit neural network: Squeeze the last bit out with admm. Cong Leng, Hao Li, Shenghuo Zhu, and Rong Jin.
- Learning accurate low-bit deep neural networks with stochastic quantization. Yinpeng Dong, Renkun Ni, Jianguo Li, Yurong Chen, Jun Zhu, and Hang Su.
- Adaptive weight compression for memory-efficient neural networks. Jong Hwan Ko, Duckhwan Kim, Taesik Na, Jaeha Kung, and Saibal Mukhopadhyay.
- Balanced quantization: An effective and efficient approach to quantized neural networks. Shu-Chang Zhou, Yu-Zhi Wang, He Wen, Qin-Yao He, and Yu-Heng Zou.
- Energy-efficient convnets through approximate computing. Bert Moons, Bert De Brabandere, Luc Van Gool, and Marian Verhelst.
- Binarized convolutional neural networks with separable filters for efficient hardware acceleration. Jeng Hau Lin, Tianwei Xing, Ritchie Zhao, Zhiru Zhang, Mani Srivastava, Zhuowen Tu, and Rajesh K. Gupta.
- Performance guaranteed network acceleration via high-order residual quantization. Wenjun Zhang XiaoKang Yang Wen Gao Zefan Li, Bingbing Ni.
- Loss-aware binarization of deep networks. Lu Hou, Quanming Yao, and James T Kwok.
- Bitnet: Bit-regularized deep neural networks. Aswin Raghavan, Mohamed Amer, and Sek Chai.
- Analytical guarantees on numerical precision of deep neural networks. Charbel Sakr, Yongjune Kim, and Naresh Shanbhag.
- Mixed low-precision deep learning inference using dynamic fixed point. Naveen Mellempudi, Abhisek Kundu, Dipankar Das, Dheevatsa Mudigere, and Bharat Kaul.
- Understanding the impact of precision quantization on the accuracy and energy of neural networks. Sherief Reda, Sherief Reda, Sherief Reda, Sherief Reda, and Sherief Reda.
- Soft-to-hard vector quantization for end-to-end learned compression of images and neural networks. Eirikur Agustsson, Fabian Mentzer, Michael Tschannen, Lukas Cavigelli, Radu Timofte, Luca Benini, and Luc Van Gool.
- Espresso: Efficient forward propagation for bcnns. Fabrizio Pedersoli, George Tzanetakis, and Andrea Tagliasacchi.
- Intra-layer nonuniform quantization of convolutional neural network. Fangxuan Sun, Jun Lin, and Zhongfeng Wang.
- Scalable compression of deep neural networks. Xing Wang and Jie Liang.
- Embedded binarized neural networks. Bradley Mcdanel, Surat Teerapittayanon, and H. T Kung.
-
-
-
-
-
-
- Predicting parameters in deep learning. Misha Denil, Babak Shakibi, Laurent Dinh, Nando de Freitas, et al.
- Learning separable filters. Roberto Rigamonti, Amos Sironi, Vincent Lepetit, and Pascal Fua.
- Speeding up convolutional neural networks with low rank expansions. Max Jaderberg, Andrea Vedaldi, and Andrew Zisserman.
- Exploiting linear structure within convolutional networks for efficient evaluation. Emily L Denton, Wojciech Zaremba, Joan Bruna, Yann LeCun, and Rob Fergus.
- Speeding-up convolutional neural networks using fine-tuned cp-decomposition. Vadim Lebedev, Yaroslav Ganin, Maksim Rakhuba, Ivan Oseledets, and Victor Lempitsky.
- Efficient and accurate approximations of nonlinear convolutional networks. Xiangyu Zhang, Jianhua Zou, Xiang Ming, Kaiming He, and Jian Sun.
- Compression of fully-connected layer in neural network by kronecker product. Shuchang Zhou and Jia-Nan Wu.
- Restructuring of deep neural network acoustic models with singular value decomposition. Jian Xue, Jinyu Li, and Yifan Gong.
- An exploration of parameter redundancy in deep networks with circulant projections. Yu Cheng, Felix X Yu, Rogerio S Feris, Sanjiv Kumar, Alok Choudhary, and Shi-Fu Chang.
- Convolutional neural networks with low-rank regularization. Cheng Tai, Tong Xiao, Yi Zhang, Xiaogang Wang, et al.
- Deep fried convnets. Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, and Ziyu Wang.
- Training cnns with low-rank filters for efficient image classification. Yani Ioannou, Duncan Robertson, Jamie Shotton, Roberto Cipolla, and Antonio Criminisi.
- Factorized convolutional neural networks. Min Wang, Baoyuan Liu, and Hassan Foroosh.
- Compression of deep convolutional neural networks for fast and low power mobile applications. Yong-Deok Kim, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu Yang, and Dongjun Shin.
- Accelerating convolutional neural networks for mobile applications. Peisong Wang and Jian Cheng.
- Decomposeme: Simplifying convnets for end-to-end learning. Jose Alvarez and Lars Petersson.
- Structured transforms for small-footprint deep learning. Vikas Sindhwani, Tara Sainath, and Sanjiv Kumar.
- Design of efficient convolutional layers using single intrachannel convolution, topological subdivisioning and spatial bottleneck structure. Min Wang, Baoyuan Liu, and Hassan Foroosh.
- Low-rank matrix factorization for deep neural network training with high-dimensional output targets. Tara N Sainath, Brian Kingsbury, Vikas Sindhwani, Ebru Arisoy, and Bhuvana Ramabhadran.
- Low precision neural networks using subband decomposition. Sek Chai, Aswin Raghavan, David Zhang, Mohamed Amer, and Tim Shields.
- Beyond filters: Compact feature map for portable deep model. Yunhe Wang, Chang Xu, Chao Xu, and Dacheng Tao.
- Theoretical properties for neural networks with weight matrices of low displacement rank. Liang Zhao, Siyu Liao, Yanzhi Wang, Jian Tang, and Bo Yuan.
- Coordinating filters for faster deep neural networks. Wei Wen, Cong Xu, Chunpeng Wu, Yandan Wang, Yiran Chen, and Hai Li.
- Ultimate tensorization: compressing convolutional and fc layers alike. Timur Garipov, Dmitry Podoprikhin, Alexander Novikov, and Dmitry Vetrov.
- Simplifying deep neural networks for neuromorphic architectures. Jaeyong Chung and Taehwan Shin.
- Network sketching: Exploiting binary structure in deep cnns. Yiwen Guo, Anbang Yao, Hao Zhao, and Yurong Chen.
- Improving efficiency in convolutional neural network with multilinear filters. Dat Thanh Tran, Alexandros Iosifidis, and Moncef Gabbouj.
- Analysis and design of convolutional networks via hierarchical tensor decompositions. Nadav Cohen, Or Sharir, Yoav Levine, Ronen Tamari, David Yakira, and Amnon Shashua.
- Structured convolution matrices for energye efficient deep learning. Rathinakumar Appuswamy, Tapan Nayak, John Arthur, Steven Esser, Paul Merolla, Jeffrey Mckinstry, Timothy Melano, Myron Flickner, and Dharmendra Modha.
- Circnn: Accelerating and compressing deep neural networks using block-circulantweight matrices. Caiwen Ding, Siyu Liao, Yanzhi Wang, Zhe Li, Ning Liu, Youwei Zhuo, Chao Wang, Xuehai Qian, Yu Bai, and Geng Yuan.
-
-
-
-
-
-
-
-
-
- Dyvedeep: Dynamic variable effort deep neural networks. Sanjay Ganapathy, Swagath Venkataramani, Balaraman Ravindran, and Anand Raghunathan.
- Spatially adaptive computation time for residual networks. Michael Figurnov, Maxwell D Collins, Yukun Zhu, Li Zhang, Jonathan Huang, Dmitry Vetrov, and Ruslan Salakhutdinov.
-
-
-
-
-
-
-
- Mobilenets: Efficient convolutional neural networks for mobile vision applications. Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam.
- Flattened convolutional neural networks for feedforward acceleration. Jonghoon Jin, Aysegul Dundar, and Eugenio Culurciello.
- Lcnn: Lookup-based convolutional neural network. Hessam Bagherinezhad, Mohammad Rastegari, and Ali Farhadi.
- Local binary convolutional neural networks. Felix Juefei-Xu, Vishnu Naresh Boddeti, and Marios Savvides.
- Squeezenet: Alexnet-level accuracy with 50x fewer parameters and< 0.5 mb model size. Forrest N Iandola, Song Han, Matthew W Moskewicz, Khalid Ashraf, William J Dally, and Kurt Keutzer.
- A compact dnn: Approaching googlenet-level accuracy of classification and domain adaptation. Chunpeng Wu, Wei Wen, Tariq Afzal, Yongmei Zhang, Yiran Chen, and Hai Li.
- Shufflenet: An extremely efficient convolutional neural network for mobile devices. Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, and Jian Sun.
- Deep simnets. Nadav Cohen, Or Sharir, and Amnon Shashua.
- Densely connected convolutional networks. Gao Huang, Zhuang Liu, Kilian Q Weinberger, and Laurens van der Maaten.
- Genetic cnn. Lingxi Xie and Alan Yuille.
- Sep-nets: Small and effective pattern networks. Zhe Li, Xiaoyu Wang, Xutao Lv, and Tianbao Yang.
- Learning the structure of deep convolutional networks. Jiashi Feng and Trevor Darrell.
- Convolutional neural fabrics. Shreyas Saxena and Jakob Verbeek.
-
-
-
-
-
- Group equivariant convolutional networks. Taco S Cohen and Max Welling.
- Doubly convolutional neural networks. Shuangfei Zhai, Yu Cheng, Weining Lu, and Zhongfei Zhang.
- Understanding and improving convolutional neural networks via concatenated rectified linear units. Wenling Shang, Kihyuk Sohn, Diogo Almeida, and Honglak Lee.
- Multi-bias non-linear activation in deep neural networks. Hongyang Li, Wanli Ouyang, and Xiaogang Wang.
- Exploiting cyclic symmetry in convolutional neural networks. Sander Dieleman, Jeffrey De Fauw, and Koray Kavukcuoglu.
- Building correlations between filters in convolutional neural networks. H. Wang, P. Chen, and S Kwong.
- Fast training of convolutional networks through ffts. Michael Mathieu, Mikael Henaff, and Yann LeCun.
- Fast algorithms for convolutional neural networks. Andrew Lavin and Scott Gray.
- Fast convolutional nets with fbfft: A gpu performance evaluation. Nicolas Vasilache, Jeff Johnson, Michael Mathieu, Soumith Chintala, Serkan Piantino, and Yann LeCun.
- Eie: efficient inference engine on compressed deep neural network. Song Han, Xingyu Liu, Huizi Mao, Jing Pu, Ardavan Pedram, Mark A Horowitz, and William J Dally.
- Cnnpack: packing convolutional neural networks in the frequency domain. Yunhe Wang, Chang Xu, Shan You, Dacheng Tao, and Chao Xu.
- Low-memory gemm-based convolution algorithms for deep neural networks. Andrew Anderson, Aravind Vasudevan, Cormac Keane, and David Gregg.
- Convolution in convolution for network in network. Y. Pang, M. Sun, X. Jiang, and X. Li.
-
-
-