On a Conjecture Regarding the Adam Optimizer

Autor: Akrout, Mohamed, Tweed, Douglas
Rok vydání: 2021
Předmět:
Druh dokumentu: Working Paper
Popis: Why does the Adam optimizer work so well in deep-learning applications? Adam's originators, Kingma and Ba, presented a mathematical argument that was meant to help explain its success, but Bock and colleagues have since reported that a key piece is missing from that argument $-$ an unproven lemma which we will call Bock's conjecture. Here we show that this conjecture is false, but we prove a modified version of it $-$ a generalization of a result of Reddi and colleagues $-$ which can take its place in analyses of Adam.
Comment: Accepted to the Optimization Algorithms and Machine Learning (ICOAML) conference
Databáze: arXiv