Layer Skip Learning using LARS variables for 39% Faster Conversion Time and Lower Bandwidth
説明
In this paper, a method for the improvement of the relationship between calculation time and recognition accuracy in deep learning is proposed. A major problem with respect to deep learning is that a large calculation time is required for higher recognition accuracy. Because of this problem, the implementation of deep learning in hardware and its application to real problems are limited. In this study, layer-wise adaptive rate scaling (LARS) variables are adopted to evaluate the necessity of the learning of each layer. When the variable of a certain convolution layer exceeds the threshold value, the learning for that layer is considered unnecessary; thus, the layer is skipped. When a layer recognized as the layer that does not require learning, only the lower layers below than that layer are learned in the next epoch. By adaptively skipping the layer, the calculation time is reduced. Furthermore, the recognition accuracy is improved. Consequently, the proposed methods accelerate the calculation time in VGG-F to achieve the highest accuracy for the top1 and top5 test accuracy by a speed up factor of 2.14, and 2.25, respectively. Moreover, the respective top1 and top5 test accuracy was improved by 3.0%, and 2.8% which obtained as the final accuracy. In addition, the operation process was reduced by approximately 39.0% and required bandwidth was reduced by 38.9%, when compared with the case of conventional full layer learning.
収録刊行物
-
- 2018 25th IEEE International Conference on Electronics, Circuits and Systems (ICECS)
-
2018 25th IEEE International Conference on Electronics, Circuits and Systems (ICECS) 673-676, 2018-12-01
IEEE