TY - GEN
T1 - LeanConvNets
T2 - Low-Cost Yet Effective Convolutional Neural Networks
AU - Ephrath, Jonathan
AU - Eliasof, Moshe
AU - Ruthotto, Lars
AU - Haber, Eldad
AU - Treister, Eran
N1 - Funding Information:
Manuscript received July 1, 2019; revised December 24, 2019 and January 23, 2020; accepted January 23, 2020. Date of publication February 10, 2020; date of current version August 10, 2020. The work of Lars Ruthotto was supported by the U.S. National Science Foundation award DMS 1751636. This work was supported by the United States - Israel Binational Science Foundation, Jerusalem, Israel under Grant 2018209. The work of Moshe Eliasof was supported by Kreitman High-Tech Scholarship. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Diana Marculescu. (Jonathan Ephrath and Moshe Eliasof contributed equally to this work.) (Corresponding author: Eran Treister.) Jonathan Ephrath, Moshe Eliasof, and Eldad Haber are with the Department of Computer Sciences, the Ben-Gurion University of the Negev, Be’er Sheva 84105, Israel (e-mail: ephrathj@post.bgu.ac.il; eliasof@post.bgu.ac.il; ehaber@eos.ubc.ca).
Publisher Copyright:
© 2007-2012 IEEE.
PY - 2020/5/1
Y1 - 2020/5/1
N2 - Convolutional Neural Networks (CNNs) have become indispensable for solving machine learning tasks in speech recognition, computer vision, and other areas that involve high-dimensional data. A CNN filters the input feature using a network containing spatial convolution operators with compactly supported stencils. In practice, the input data and the hidden features consist of a large number of channels, which in most CNNs are fully coupled by the convolution operators. This coupling leads to immense computational cost in the training and prediction phase. In this article, we introduce LeanConvNets that are derived by sparsifying fully-coupled operators in existing CNNs. Our goal is to improve the efficiency of CNNs by reducing the number of weights, floating point operations and latency times, with minimal loss of accuracy. Our lean convolution operators involve tuning parameters that controls the trade-off between the network's accuracy and computational costs. These convolutions can be used in a wide range of existing networks, and we exemplify their use in residual networks (ResNets). Using a range of benchmark problems from image classification and semantic segmentation, we demonstrate that the resulting LeanConvNet's accuracy is close to state-of-the-art networks while being computationally less expensive. In our tests, the lean versions of ResNet in most cases outperform comparable reduced architectures such as MobileNets and ShuffleNets.
AB - Convolutional Neural Networks (CNNs) have become indispensable for solving machine learning tasks in speech recognition, computer vision, and other areas that involve high-dimensional data. A CNN filters the input feature using a network containing spatial convolution operators with compactly supported stencils. In practice, the input data and the hidden features consist of a large number of channels, which in most CNNs are fully coupled by the convolution operators. This coupling leads to immense computational cost in the training and prediction phase. In this article, we introduce LeanConvNets that are derived by sparsifying fully-coupled operators in existing CNNs. Our goal is to improve the efficiency of CNNs by reducing the number of weights, floating point operations and latency times, with minimal loss of accuracy. Our lean convolution operators involve tuning parameters that controls the trade-off between the network's accuracy and computational costs. These convolutions can be used in a wide range of existing networks, and we exemplify their use in residual networks (ResNets). Using a range of benchmark problems from image classification and semantic segmentation, we demonstrate that the resulting LeanConvNet's accuracy is close to state-of-the-art networks while being computationally less expensive. In our tests, the lean versions of ResNet in most cases outperform comparable reduced architectures such as MobileNets and ShuffleNets.
KW - Moshe: Computer vision
KW - deep convolutional neural networks
KW - intelligent systems
KW - machine learning
UR - http://www.scopus.com/inward/record.url?scp=85079465448&partnerID=8YFLogxK
U2 - 10.1109/JSTSP.2020.2972775
DO - 10.1109/JSTSP.2020.2972775
M3 - Conference contribution
AN - SCOPUS:85079465448
VL - 14
T3 - IEEE Journal on Selected Topics in Signal Processing
SP - 894
EP - 904
BT - 36th International Conference on Machine Learning Workshop (ICML), Long Beach, CA, USA, 2019
ER -