Adam optimization algorithm DL; Machine Learning
noun phrase
Definition: A stochastic gradient-based optimization algorithm that uses adaptive estimates of first-order and second-order moments of gradients to update parameters efficiently during training [Keras Documentation].
Example in context: “Then we use the Adam optimization algorithm to update the parameters of the network.” [Liu et al. 2022]
Synonyms: Adam; adaptive moment estimation
Related terms: SGD; RMSProp; momentum