A Comparative Analysis of Gradient Descent-Based Optimization Algorithms on Convolutional Neural Networks

dc.contributor.authorDogo, E. M.
dc.contributor.authorAfolabi, O. J.
dc.contributor.authorNwulu, N. I.
dc.contributor.authorTwala, B.
dc.contributor.authorAigbavboa, C. O.
dc.date.accessioned2025-04-25T12:06:35Z
dc.date.issued2018
dc.description2018 International Conference on Computational Techniques, Electronics and Mechanical Systems (CTEMS), Belgaum, India, 2018, pp. 92-99
dc.description.abstractIn this paper, we perform a comparative evaluation of seven most commonly used first-order stochastic gradient-based optimization techniques in a simple Convolutional Neural Network (ConvNet) architectural setup. The investigated techniques are the Stochastic Gradient Descent (SGD), with vanilla (vSGD), with momentum (SGDm), with momentum and nesterov (SGDm+n)), Root Mean Square Propagation (RMSProp), Adaptive Moment Estimation (Adam), Adaptive Gradient (AdaGrad), Adaptive Delta (AdaDelta), Adaptive moment estimation Extension based on infinity norm (Adamax) and Nesterov-accelerated Adaptive Moment Estimation (Nadam). We trained the model and evaluated the optimization techniques in terms of convergence speed, accuracy and loss function using three randomly selected publicly available image classification datasets. The overall experimental results obtained show Nadam achieved better performance across the three datasets in comparison to the other optimization techniques, while AdaDelta performed the worst.
dc.identifier.otherdoi: 10.1109/CTEMS.2018.8769211.
dc.identifier.urihttp://repository.futminna.edu.ng:4000/handle/123456789/1007
dc.language.isoen
dc.publisherIEEE
dc.subjectOptimization
dc.subjectTraining
dc.subjectDeep learning
dc.subjectNeural networks
dc.subjectStochastic processes
dc.subjectConvergence
dc.subjectClassification algorithms
dc.subjectArtificial Intelligence
dc.subjectoptimizers
dc.subjectperformance measures
dc.subjectdeep learning
dc.subjectstochastic gradient descent
dc.titleA Comparative Analysis of Gradient Descent-Based Optimization Algorithms on Convolutional Neural Networks
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
A Comparative Analysis of Gradient Descent-Based Optimization Algorithms_CNN.pdf
Size:
58.48 KB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed to upon submission
Description: