A Comparative Exploration of Activation Functions for Image Classification in Convolutional Neural Networks

Faiza Makhdoom*, Jamshaid Ul Rahman**
* Abdus Salam School of Mathematical Sciences, Government College University, Lahore, Pakistan.
** School of Mathematical Sciences, Jiangsu University, Zhenjiang, China.
Periodicity:January - June'2024
DOI : https://doi.org/10.26634/jaim.2.1.20225

Abstract

Activation functions play a crucial role in enabling neural networks to carry out tasks with increased flexibility by introducing non-linearity. The selection of appropriate activation functions becomes even more crucial, especially in the context of deeper networks where the objective is to learn more intricate patterns. Among various deep learning tools, Convolutional Neural Networks (CNNs) stand out for their exceptional ability to learn complex visual patterns. In practice, ReLu is commonly employed in convolutional layers of CNNs, yet other activation functions like Swish can demonstrate superior training performance while maintaining good testing accuracy on different datasets. This paper presents an optimally refined strategy for deep learning-based image classification tasks by incorporating CNNs with advanced activation functions and an adjustable setting of layers. A thorough analysis has been conducted to support the effectiveness of various activation functions when coupled with the favorable softmax loss, rendering them suitable for ensuring a stable training process. The results obtained on the CIFAR-10 dataset demonstrate the favorability and stability of the adopted strategy throughout the training process.

Keywords

Activation Functions, Image Classification, Convolutional Neural Network,DeepLearning, Machine Intelligence.

How to Cite this Article?

Makhdoom, F., and Rahman, J. U. (2024). A Comparative Exploration of Activation Functions for Image Classification in Convolutional Neural Networks. i-manager’s Journal on Artificial Intelligence & Machine Learning, 2(1), 9-17. https://doi.org/10.26634/jaim.2.1.20225

References

[2]. Alpaydin, E. (2020). Introduction to Machine Learning. MIT press.
[3]. Bonnell, J. A. (2011). Implementation of a New Sigmoid Function in Back propagation Neural Networks (Master thesis, East Tennessee State University).
[4]. Chaithanya, B. N., Swasthika Jain, T. J., Usha Ruby, A., & Parveen, A. (2021). An approach to categorize chest X- ray images using sparse categorical cross entropy. Indonesian Journal of Electrical Engineering and Computer Science, 24(3), 1700-1710.
[5]. Covington, P., Adams, J., & Sargin, E. (2016, September). Deep neural networks for youtube recommendations. In Proceedings of the 10th ACM Conference on Recommender Systems, (pp. 191-198).
[6]. da Silva, I. N., Hernane Spatti, D., Andrade Flauzino, R., Liboni, L. H. B., dos Reis Alves, S. F., da Silva, I. N., & dos Reis Alves, S. F. (2017). Multilayer perceptron networks. Artificial Neural Networks: A Practical Course, (pp. 55-115).
[10]. Deng, J., Guo, J., Xue, N., & Zafeiriou, S. (2019). Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, (pp. 4690- 4699).
[14]. Gulli, A., & Pal, S. (2017). Deep Learning with Keras. Packt Publishing Ltd.
[24]. Liu, W., Wen, Y., Yu, Z., Li, M., Raj, B., & Song, L. (2017). Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 212-220).
[26]. Liu, Y., Gao, Y., & Yin, W. (2020). An improved analysis of stochastic gradient descent with momentum. Advances in Neural Information Processing Systems, 33, 18261-18271.
[37]. Sharma, S., Sharma, S., & Athaiya, A. (2017). Activation functions in neural networks. International Journal of Engineering Applied Sciences and Technology, 4(12), 310-316.
[38]. Tan, M., & Le, Q. (2019, May). Efficientnet: Rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning, (pp. 6105-6114). PMLR.
[40]. Wang, H., Wang, Y., Zhou, Z., Ji, X., Gong, D., Zhou, J., & Liu, W. (2018). Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 5265-5274).
[44]. Zhang, C., Benz, P., Argaw, D. M., Lee, S., Kim, J., Rameau, F., & Kweon, I. S. (2021). Resnet or densenet? Introducing dense shortcuts to resnet. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, (pp. 3550-3559).
If you have access to this article please login to view the article or kindly login to purchase the article

Purchase Instant Access

Single Article

North Americas,UK,
Middle East,Europe
India Rest of world
USD EUR INR USD-ROW
Online 15 15

Options for accessing this content:
  • If you would like institutional access to this content, please recommend the title to your librarian.
    Library Recommendation Form
  • If you already have i-manager's user account: Login above and proceed to purchase the article.
  • New Users: Please register, then proceed to purchase the article.