Performance Analysis of Convolutional Neural Networks for Image Classification with Appropriate Optimizers

Sana Danish*, Jamshaid Ul Rahman**, Gulfam Haider***
* Abdus Salam School of Mathematical Sciences, GC University, Lahore, Pakistan.
** School of Mathematical Sciences, Jiangsu University 301 Xuefu road, Zhenjiang, China.
*** FAST University of Computer and Emerging Sciences (NUCES), Chiniot Faisalabad (CFD), Campus, Pakistan.
Periodicity:January - June'2023
DOI : https://doi.org/10.26634/jmat.12.1.19398

Abstract

Optimizers in Convolutional Neural Networks play an important role in many advanced deep learning models. Studies on advanced optimizers and modifications of existing optimizers continue to hold significant importance in the study of machine tools and algorithms. There are a number of studies to defend and the selection of these optimizers illustrate some of the challenges on the effectiveness of these optimizers. Comprehensive analysis on the optimizers and alteration with famous activation function Rectified Linear Unit (ReLU) offered to protect effectiveness. Significance is determined based on the adjustment with the original Softmax and ReLU. Experiments were performed with Adam, Root Mean Squared Propagation (RMSprop), Adaptive Learning Rate Method (Adadelta), Adaptive Gradient Algorithm (Adagrad) and Stochastic Gradient Descent (SGD) to examine the performance of Convolutional Neural Networks for image classification using the Canadian Institute for Advanced Research dataset (CIFAR-10).

Keywords

Adam, Convolutional Neural Network, Optimizers, RMSprop, Stochastic Gradient Descent.

How to Cite this Article?

Danish, S., Rahman, J. U., and Haider, G. (2023). Performance Analysis of Convolutional Neural Networks for Image Classification with Appropriate Optimizers. i-manager’s Journal on Mathematics, 12(1), 1-8. https://doi.org/10.26634/jmat.12.1.19398

References

[6]. Cui, X., Zhang, W., Tüske, Z., & Picheny, M. (2018). Evolutionary stochastic gradient descent for optimization of deep neural networks. Advances in Neural Information Processing Systems, 31.
[8]. Hayou, S., Doucet, A., & Rousseau, J. (2019, May). On the impact of the activation function on deep neural networks training. In International Conference on Machine Learning (pp. 2672-2680). PMLR.
[15]. Li, Y., & Yuan, Y. (2017). Convergence analysis of two-layer neural networks with relu activation. Advances in Neural Information Processing Systems, 30.
[17]. Lydia, A., & Francis, S. (2019). Adagrad-an optimizer for stochastic gradient descent. International Journal of Information and Computing Science, 6(5), 566-568.
[25]. Shinde, P. P., & Shah, S. (2018, August). A review of machine learning and deep learning applications. In 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA) (pp. 1-6). IEEE.
[31]. Van Rossum, G., & Drake Jr, F. L. (1995). Python Tutorial (Vol. 620). Centrum voor Wiskunde en Informatica, Amsterdam, The Netherlands.
[34]. Xu, J., Zhang, Z., Friedman, T., Liang, Y., & Broeck, G. (2018, July). A semantic loss function for deep learning with symbolic knowledge. In International Conference on Machine Learning (pp. 5502-5511). PMLR.
If you have access to this article please login to view the article or kindly login to purchase the article

Purchase Instant Access

Single Article

North Americas,UK,
Middle East,Europe
India Rest of world
USD EUR INR USD-ROW
Online 15 15

Options for accessing this content:
  • If you would like institutional access to this content, please recommend the title to your librarian.
    Library Recommendation Form
  • If you already have i-manager's user account: Login above and proceed to purchase the article.
  • New Users: Please register, then proceed to purchase the article.