SPECIFICS OF THE LEARNING ERROR DEPENDENCE OF MULTILAYERED NEURAL NETWORKS FROM THE ACTIVATION FUNCTION DURING THE PROCESS OF PRINTED DIGITS IDENTIFICATION

Serhiy Sveleba, Ivan Katerynchuk, Ivan Kuno, Ostap Semotiuk, Yaroslav Shmyhelskyy, N. Sveleba

Abstract


In this paper, we investigated the learning process during an identification of printed digits from the type of an activation function. The study of the activation function type and the number of iterations in the learning process of the neural system was carried out using the Fourier spectra analysis of the learning error function and branching diagrams. For this purpose, a program for the multilayer neural network was developed in the Python software, which involves setting the number of hidden layers as well as the number of neurons inside them and changes in the learning rate. The learning rate was considered as a constant, and its optimal value, where the best learning rate is observed, was determined. To analyze the learning rate effect on the educational process, we used a logistic function describing the frequency doubling process. It is shown that the learning error function is characterized by bifurcation processes leading to a chaotic state when η>0.8. The optimal learning rate value that determines the emergence of the doubling process of the local minima number is determined. It was found that the sigmoidal activation function (as compared to the activation functions ReLU and hyperbolic tangent) best satisfies the learning process of the three-layer neural network for recognizing digits, given an array of 4x7 zeros and ones. Compared to other activation functions, there is an insignificant change in the learning error during the transition from one digit to another. It is shown that an increase in the number of hidden layers does not lead to a sharp increase during the learning error. An increase in the learning iterations number is accompanied by the appearance of periodic dependences of the logistic function value of the learning rate, the period of which is a variable of the number of iterations and the learning rate. Using Fourier spectra of the error function from the learning rate value, it can be argued that an increase in the number of iterations leads to an increase in the number of harmonics, which eventually leads to the appearance of a chaotic state of the neural network.

Keywords: Multilayered neural network, activation function, optimal learning rate, digital identification.


Full Text:

PDF

References


[1] S. Mogilnyj Machine learning with the use of microcomputers: teaching book editors: O. Lisovyj and others. - K., 2019. - 226 p.

[2] I. Goodfellow, Y. Bengio, A. Courville. Deep Learning, 2016 URL: http://www.deeplearningbook.org

[3] S. Sveleba, I. Katerynchuk, I. Kuno, I. Karpa, O. Semotyuk, Ya. Shmygelsky, N. Sveleba, V. Kuno. Chaotic states of a multilayer neural network, Electronics and information technologies. 2021. Issue 16. P. 20-35.

[4] X.-S. Wei. Must Know Tips/Tricks in Deep Neural Networks URL: http://www.lamda.nju.edu.cn/weixs/project/CNNTricks/CNNTricks.html

[5] J. Brownlee. A Gentle Introduction to the Rectified Linear Unit (ReLU) URL: https://machinelearningmastery.com/rectified-linear-activation-function-for-deep-learning-neural-networks/

[6] M. Nielsen Neural Networks and Deep Learning Chapter 1: Using neural nets to recognize handwritten digits URL: http://neuralnetworksanddeeplearning.com/chap1.html

[7] Yu. Olenych, S. Sveleba, I. Katerynchuk, I. Kunio, I. Karpa. Features of deep studyneural network. 2019. URL: https://openreviewhub.org/lea/paper-2019/features-deep-study-neural-network#

[8] Yu. Taranenko Information entropy of chaos URL: https://habr.com/ru/post/447874/




DOI: http://dx.doi.org/10.30970/eli.17.4

Refbacks

  • There are currently no refbacks.