Browse > Article
http://dx.doi.org/10.3745/KTSDE.2020.9.10.303

A Performance Comparison of Super Resolution Model with Different Activation Functions  

Yoo, Youngjun (국민대학교 컴퓨터공학과)
Kim, Daehee (국민대학교 컴퓨터공학과)
Lee, Jaekoo (국민대학교 SW학부)
Publication Information
KIPS Transactions on Software and Data Engineering / v.9, no.10, 2020 , pp. 303-308 More about this Journal
Abstract
The ReLU(Rectified Linear Unit) function has been dominantly used as a standard activation function in most deep artificial neural network models since it was proposed. Later, Leaky ReLU, Swish, and Mish activation functions were presented to replace ReLU, which showed improved performance over existing ReLU function in image classification task. Therefore, we recognized the need to experiment with whether performance improvements could be achieved by replacing the RELU with other activation functions in the super resolution task. In this paper, the performance was compared by changing the activation functions in EDSR model, which showed stable performance in the super resolution task. As a result, in experiments conducted with changing the activation function of EDSR, when the resolution was converted to double, the existing activation function, ReLU, showed similar or higher performance than the other activation functions used in the experiment. When the resolution was converted to four times, Leaky ReLU and Swish function showed slightly improved performance over ReLU. PSNR and SSIM, which can quantitatively evaluate the quality of images, were able to identify average performance improvements of 0.06%, 0.05% when using Leaky ReLU, and average performance improvements of 0.06% and 0.03% when using Swish. When the resolution is converted to eight times, the Mish function shows a slight average performance improvement over the ReLU. Using Mish, PSNR and SSIM were able to identify an average of 0.06% and 0.02% performance improvement over the RELU. In conclusion, Leaky ReLU and Swish showed improved performance compared to ReLU for super resolution that converts resolution four times and Mish showed improved performance compared to ReLU for super resolution that converts resolution eight times. In future study, we should conduct comparative experiments to replace activation functions with Leaky ReLU, Swish and Mish to improve performance in other super resolution models.
Keywords
Super Resolution; Performance Comparison; EDSR; Activation Function;
Citations & Related Records
연도 인용수 순위
  • Reference
1 R. Timofte, E. Agustsson, L. Van Gool, M.-H. Yang, L. Zhang, B. Lim, S. Son, H. W. Kim, et al., "NTIRE 2017 Challenge on Single Image Super-Resolution: Methods and Results," In Conference on Computer Vision and Pattern Recognition, 2017, 1,2,4,6,7,8.
2 Saeed Anwar, Salman Khan, and Nick Barnes, "A Deep Journey into Super-resolution: A Survey," arXiv:1904.07523, 1, 2019.
3 B. Lim, S. Son, H. Kim, S. Nah, and K. M. Lee, "Enhanced Deep Residual Networks for Single Image Super-Resolution," In Conference on Computer Vision and Pattern Recognition, 2017, 1.
4 C. Ledig, L. Thesis, F. Huszar, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani, J. Totz, Z. Wang, et al, "Photo-realistic Single Image Super-resolution using a Generative Adversarial network," arXiv:1609.040802,1,2,3,4,5,6,7, 2017.
5 K. He, X. Zhang, S. Ren, and J. Sun, "Deep Residual Learning for Image Recognition," In Conference on Computer Vision and Pattern Recognition, 2016, 3.
6 Nair, Vinod and Hinton, Geoffrey E, "Rectified Linear units Improve Restricted Boltzmann Machines," In International Conference on Machine Learning, 2010, pp.807-814.
7 Bing Xu, Naiyan Wang, Tianqi Chen, Mu Li, "Empirical Evaluation of Rectified Activations in Convolutional Network," arXiv:1505.00853, 1, 2015.
8 R. Zeyde, M. Elad, and M. Protter, "On Single Image Scale-up using Sparse-representations," In Proceedings of the International Conference on Curves and Surfaces, 2010, 2,4.
9 Diganta Misra, "Mish: A Self Regularized Non-Monotonic Neu ral Activation Function," arXiv:1908.08681, 1, 2019.
10 M. Bevilacqua, A. Roumy, C. Guillemot, and M. L. Alberti-Morel, "Low-complexity Single-image Super-resolution Based on Nonnegative Neighbor Embedding," In British Machine Vision Conference, 2012, 2,4.
11 D. Martin, C. Fowlkes, D. Tal, and J. Malik, "A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics," In International Conference on Computer Vision, 2001, 4.
12 Prajit Ramachandran, Barret Zoph, and Quoc V. Le, "Swish: a Self-gated Activation Function," arXiv:1710.05941 7, 1, 2017.
13 J.-B. Huang, A. Singh, and N. Ahuja, "Single Image Super-resolution from Transformed Self-exemplars," In Conference on Computer Vision and Pattern Recognition, 2015, 2,4,6.