How to choose an optimizer for your Neural Network?

Cross Validated Asked by White on January 1, 2022

My question boils down to this : Which optimizer should I use to train my neural network.

I understand this question depends on the problem.

However, for instance it seems that ADAM optimizer generally outperforms SGD. So maybe we can find some rationality in the choice of optimizers (and their learning rate) ?

Maybe it won’t be mathematically correct, but maybe some of you came accross a standard method for choosing an optimizer ?

One Answer

There is no science behind which optimiser to use in a model. It mostly boils down to heuristics and type of model in use.

ADAM is not generally better than SGD: This blog post reviews an article about SGD being a better optimiser than ADAM.

There are pros and cons of each optimiser and no optimiser works best This link compares different gradient descent optimisers.

Answered by Vivek on January 1, 2022

Add your own answers!

Ask a Question

Get help from others!

© 2024 All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP