Browse > Article
http://dx.doi.org/10.6109/jkiice.2022.26.1.49

Calculating Data and Artificial Neural Network Capability  

Yi, Dokkyun (Seongsan Liberal Arts College, Daegu University)
Park, Jieun (Seongsan Liberal Arts College, Daegu University)
Abstract
Recently, various uses of artificial intelligence have been made possible through the deep artificial neural network structure of machine learning, demonstrating human-like capabilities. Unfortunately, the deep structure of the artificial neural network has not yet been accurately interpreted. This part is acting as anxiety and rejection of artificial intelligence. Among these problems, we solve the capability part of artificial neural networks. Calculate the size of the artificial neural network structure and calculate the size of data that the artificial neural network can process. The calculation method uses the group method used in mathematics to calculate the size of data and artificial neural networks using an order that can know the structure and size of the group. Through this, it is possible to know the capabilities of artificial neural networks, and to relieve anxiety about artificial intelligence. The size of the data and the deep artificial neural network are calculated and verified through numerical experiments.
Keywords
Explainable artificial intelligence; Artificial intelligence capabilities; Data size; Artificial neural network structure; Analysis;
Citations & Related Records
연도 인용수 순위
  • Reference
1 C. M. Bishop, Pattern Recognition and Machine Learning, New York, NY: Springer-Verlag, 2006.
2 H. Zulkifli, Understanding Learning Rates and How It Improves Performance in Deep Learning. Towards Data Science [Internet]. Available: https://towardsdatascience.com/understanding-learning-rates-and-how-it-improves-performance-in-deep-learning-d0d4059c1c10.
3 A. Krizhevsky, H. Suskever, and G. E. Hinton, "ImageNet Classification with Deep Convolutional Neural Networks," Communications of the ACM, vol. 60, no. 6, pp. 84-90, Jun. 2017.   DOI
4 A. Ng. The most frequently themes(AItimes) [Internet]. Available: http://www.aitimes.com/news/articleView.html?idxno=131542.
5 S. Lau. Learning Rate Schedules and Adaptive Learning Rate Methods for Deep Learning. Towards Data Science [Internet]. Available: https://towardsdatascience.com/learning-rate-schedules-and-adaptive-learning-rate-methods-for-deep-learning-2c8f433990d1.
6 G. Aurelien, Hands-On Machine Learning with Scikit-Learn and TensorFlow, O'Reilly Media Inc, 2017.
7 K. He, X. Zhang, S. Ren, and J. Sun, "Deep Residual Learning for Image Recognition," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas: NV, pp. 770-778, 2016.
8 I. Goodfellow, Y. Bengio, and A. Courville, "Regularization for deep learning," in Deep Learning, Cambridge : MIT Press, 2016.
9 R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, "Grad-CAM: Visual explanations from deep networks via gradient-based localization," in Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 618-626, 2017.
10 M. W. Huang, C. W. Chen, W. C. Lin, S. W. Ke, and C. F. Tsai, "Svm and svm ensembles in breast cancer prediction," PLoS ONE, vol. 12, no. 1, pp. 1-14, Jan. 2017.
11 B. Neyshabur, S. Bhojanapalli, D. McAllester, and N. Srebro, "Exploring Generalization in Deep Learning," in the 31st Conference on Neural Information Processing Systems (NIPS), Long Beach: CA, 2017.
12 Machine Learning Mastery. A Gentle Introduction to the Rectified Linear Unit (ReLU) [Internet]. Available: https://machinelearningmastery.com/rectified-linear-activation-function-for-deep-learning-neural-networks/?nowprocket=1.
13 D. X. Zhou, "Deep distributed convolutional neural networks: Universality," Analysis and Applications, vol. 16, no. 6, pp. 895-919, 2018.   DOI
14 L. Serge, Algebra, New York, NY: Springer-Verlag, 2002.
15 M. Kirby, Geometric Data Analysis, New York, NY: Wiley-Interscience, 2001.
16 Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, "Gradient-based learning applied to document recognition," Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, 1998.   DOI
17 G. Montavon, S. Bach, A. Binder, W. Samek, and K. Muller, "Explaining nonlinear classification decisions with deep taylor decomposition," Pattern Recognition, vol. 65, pp. 211-222, May. 2017.   DOI
18 M. J. Kochenderfer and T. A. Wheeler, Algorithms for Optimization, Cambridge: The MIT Press, 2019.