menu
techminis

A naukri.com initiative

google-web-stories
Home

>

Deep Learning News

>

Effectiven...
source image

Medium

4w

read

337

img
dot

Image Credit: Medium

Effectiveness of Stochastic Gradient Descent in Multilayer Perceptrons

  • Gradient descent is crucial in decreasing the loss function of neural networks for accuracy by finding the gradient of the loss function with respect to the weights.
  • Stochastic Gradient Descent (SGD) in batches of training data makes the descent less accurate yet more efficient compared to other optimizers like Adam, which is an adaptive learning rate optimizer.
  • In testing the effectiveness on MNIST handwritten digits dataset, SGD shows roughly 91% accuracy in 43 seconds, outperforming non-SGD models with an average accuracy of 87.47% in 54 seconds.
  • Despite being faster than regular gradient descent, SGD falls short compared to optimizers like Adam and RMSprop, which show higher accuracies and similar training times.

Read Full Article

like

20 Likes

For uninterrupted reading, download the app