Adaptive gradient descent for convex and non-convex stochastic optimization

dc.bibliographicCitation.volume2655
dc.contributor.authorOgaltsov, Aleksandr
dc.contributor.authorDvinskikh, Darina
dc.contributor.authorDvurechensky, Pavel
dc.contributor.authorGasnikov, Alexander
dc.contributor.authorSpokoiny, Vladimir
dc.date.accessioned2022-06-23T14:49:32Z
dc.date.available2022-06-23T14:49:32Z
dc.date.issued2019
dc.description.abstractIn this paper we propose several adaptive gradient methods for stochastic optimization. Our methods are based on Armijo-type line search and they simultaneously adapt to the unknown Lipschitz constant of the gradient and variance of the stochastic approximation for the gradient. We consider an accelerated gradient descent for convex problems and gradient descent for non-convex problems. In the experiments we demonstrate superiority of our methods to existing adaptive methods, e.g. AdaGrad and Adam.eng
dc.description.versionpublishedVersioneng
dc.identifier.urihttps://oa.tib.eu/renate/handle/123456789/9239
dc.identifier.urihttps://doi.org/10.34657/8277
dc.language.isoeng
dc.publisherBerlin : Weierstraß-Institut für Angewandte Analysis und Stochastik
dc.relation.doihttps://doi.org/10.20347/WIAS.PREPRINT.2655
dc.relation.hasversionhttps://doi.org/10.1016/j.ifacol.2020.12.2284
dc.relation.ispartofseriesPreprint / Weierstraß-Institut für Angewandte Analysis und Stochastik ; 2655
dc.relation.issn2198-5855
dc.rights.licenseThis document may be downloaded, read, stored and printed for your own use within the limits of § 53 UrhG but it may not be distributed via the internet or passed on to external parties.eng
dc.rights.licenseDieses Dokument darf im Rahmen von § 53 UrhG zum eigenen Gebrauch kostenfrei heruntergeladen, gelesen, gespeichert und ausgedruckt, aber nicht im Internet bereitgestellt oder an Außenstehende weitergegeben werden.ger
dc.subjectConvex and non-convex optimizationeng
dc.subjectstochastic optimizationeng
dc.subjectfirst-order methodeng
dc.subjectadaptive methodeng
dc.subjectgradient descenteng
dc.subjectcomplexity boundseng
dc.subjectmini-batcheng
dc.subject.ddc510
dc.titleAdaptive gradient descent for convex and non-convex stochastic optimizationeng
dc.typereporteng
dc.typeTexteng
dcterms.bibliographicCitation.journalTitlePreprint / Weierstraß-Institut für Angewandte Analysis und Stochastik
dcterms.extent17 S.
tib.accessRightsopenAccess
wgl.contributorWIAS
wgl.subjectMathematik
wgl.typeReport / Forschungsbericht / Arbeitspapier
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
wias_preprints_2655.pdf
Size:
550.04 KB
Format:
Adobe Portable Document Format
Description:
Collections