Performance of Different Activation Functions in Two Hidden Layer Handwritten Digit Recognition Neural Network

Authors

  • Yew Joe Leong Universiti Tun Hussein Onn Author
  • Siaw Chong Lee Universiti Tun Hussein Onn Malaysia Author

Keywords:

neural network, handwritten digit recognition, activation function, hyperbolic tangent function, sigmoid function, rectified linear unit function

Abstract

This research delves into the realm of handwritten digits recognition using neural
networks, with a primary objective of comparing the performance of three distinct
activation functions—tanh, sigmoid, and ReLU. Initial investigations focus on
identifying optimal parameters, involving weight initialization method, the number of
nodes in hidden layers, and learning rate. Through systematic experimentation, the
study discerns the most effective configurations for these parameters. Subsequently,
the chosen configurations are employed to evaluate the performance of the neural
network with each activation function. Results indicate that ReLU outperforms both
sigmoid and tanh, establishing itself as the most effective activation function for the
given task. These findings provide valuable insights for optimizing neural network
architectures in handwritten digit recognition applications.

Downloads

Download data is not yet available.

Downloads

Published

17-12-2024

Issue

Section

Mathematics

How to Cite

Leong, Y. J., & Lee, S. C. (2024). Performance of Different Activation Functions in Two Hidden Layer Handwritten Digit Recognition Neural Network. Enhanced Knowledge in Sciences and Technology, 4(2), 171-180. https://publisher.uthm.edu.my/periodicals/index.php/ekst/article/view/14307