Shift operation is an efficient alternative over depthwise separable
convolution. However, it is still bottlenecked by its implementation manner,
namely memory movement. To put this direction forward, a new and novel basic
component named Sparse Shift Layer (SSL) is introduced in this paper to
construct efficient convolutional neural networks. In this family of
architectures, the basic block is only composed by 1x1 convolutional layers
with only a few shift operations applied to the intermediate feature maps. To
make this idea feasible, we introduce shift operation penalty during
optimization and further propose a quantization-aware shift learning method to
impose the learned displacement more friendly for inference. Extensive ablation
studies indicate that only a few shift operations are sufficient to provide
spatial information communication. Furthermore, to maximize the role of SSL, we
redesign an improved network architecture to Fully Exploit the limited capacity
of neural Network (FE-Net). Equipped with SSL, this network can achieve 75.0%
top-1 accuracy on ImageNet with only 563M M-Adds. It surpasses other
counterparts constructed by depthwise separable convolution and the networks
searched by NAS in terms of accuracy and practical speed.
Description
All You Need is a Few Shifts: Designing Efficient Convolutional Neural Networks for Image Classification
%0 Generic
%1 chen2019shifts
%A Chen, Weijie
%A Xie, Di
%A Zhang, Yuan
%A Pu, Shiliang
%D 2019
%K arch backbone classification mobilenet regularization
%T All You Need is a Few Shifts: Designing Efficient Convolutional Neural
Networks for Image Classification
%U http://arxiv.org/abs/1903.05285
%X Shift operation is an efficient alternative over depthwise separable
convolution. However, it is still bottlenecked by its implementation manner,
namely memory movement. To put this direction forward, a new and novel basic
component named Sparse Shift Layer (SSL) is introduced in this paper to
construct efficient convolutional neural networks. In this family of
architectures, the basic block is only composed by 1x1 convolutional layers
with only a few shift operations applied to the intermediate feature maps. To
make this idea feasible, we introduce shift operation penalty during
optimization and further propose a quantization-aware shift learning method to
impose the learned displacement more friendly for inference. Extensive ablation
studies indicate that only a few shift operations are sufficient to provide
spatial information communication. Furthermore, to maximize the role of SSL, we
redesign an improved network architecture to Fully Exploit the limited capacity
of neural Network (FE-Net). Equipped with SSL, this network can achieve 75.0%
top-1 accuracy on ImageNet with only 563M M-Adds. It surpasses other
counterparts constructed by depthwise separable convolution and the networks
searched by NAS in terms of accuracy and practical speed.
@misc{chen2019shifts,
abstract = {Shift operation is an efficient alternative over depthwise separable
convolution. However, it is still bottlenecked by its implementation manner,
namely memory movement. To put this direction forward, a new and novel basic
component named Sparse Shift Layer (SSL) is introduced in this paper to
construct efficient convolutional neural networks. In this family of
architectures, the basic block is only composed by 1x1 convolutional layers
with only a few shift operations applied to the intermediate feature maps. To
make this idea feasible, we introduce shift operation penalty during
optimization and further propose a quantization-aware shift learning method to
impose the learned displacement more friendly for inference. Extensive ablation
studies indicate that only a few shift operations are sufficient to provide
spatial information communication. Furthermore, to maximize the role of SSL, we
redesign an improved network architecture to Fully Exploit the limited capacity
of neural Network (FE-Net). Equipped with SSL, this network can achieve 75.0%
top-1 accuracy on ImageNet with only 563M M-Adds. It surpasses other
counterparts constructed by depthwise separable convolution and the networks
searched by NAS in terms of accuracy and practical speed.},
added-at = {2019-04-05T23:06:01.000+0200},
author = {Chen, Weijie and Xie, Di and Zhang, Yuan and Pu, Shiliang},
biburl = {https://www.bibsonomy.org/bibtex/27c4a8f5cb86db2438621437efe703dab/nmatsuk},
description = {All You Need is a Few Shifts: Designing Efficient Convolutional Neural Networks for Image Classification},
interhash = {2669176b84632439d16f58696fe79251},
intrahash = {7c4a8f5cb86db2438621437efe703dab},
keywords = {arch backbone classification mobilenet regularization},
note = {cite arxiv:1903.05285Comment: CVPR2019},
timestamp = {2019-04-05T23:06:01.000+0200},
title = {All You Need is a Few Shifts: Designing Efficient Convolutional Neural
Networks for Image Classification},
url = {http://arxiv.org/abs/1903.05285},
year = 2019
}