Comparison of Regularization and Optimization Methods for Process Recognition with Use of Deep Neural Network
PDF

Keywords

regularization, optimization, gradient descent, learning rate decay, machine learning

How to Cite

Képešiová, Z., & Kozák, Štefan. (2023). Comparison of Regularization and Optimization Methods for Process Recognition with Use of Deep Neural Network. Information Technology Applications, 8(2), 3–14. Retrieved from https://www.itajournal.com/index.php/ita/article/view/49

Abstract

This paper brings the topic of comparison of optimization techniques using gradient descent, gradient descent with momentum, Adam and learning rate decay in combination with previous optimization algorithms for face recognition using deep neural network. This paper compares several settings of these techniques as well as techniques among themselves with combination of regularization techniques varying in using no regularization, L2 regularization and dropout to successfully recognize the sex of a person captured. The result of the evaluation is taken in a matter of successfully recognized face rate for training data and test data, cost and deep neural network learning time.

PDF
Creative Commons License

This work is licensed under a Creative Commons Attribution 4.0 International License.

Copyright (c) 2019 International Journal of Information Technology Applications