Open Access System for Information Sharing

Login Library

 

Article
Cited 2 time in webofscience Cited 2 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.authorBUM, JUN KIM-
dc.contributor.authorHYEYEON, CHOI-
dc.contributor.authorJANG, HYEONAH-
dc.contributor.authorKIM, SANG WOO-
dc.date.accessioned2023-02-22T01:20:44Z-
dc.date.available2023-02-22T01:20:44Z-
dc.date.created2023-02-21-
dc.date.issued2022-10-
dc.identifier.issn0924-669X-
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/115387-
dc.description.abstractDeep neural network optimization is challenging. Large gradients in their chaotic loss landscape lead to unstable behavior during gradient descent. In this paper, we investigate a stable gradient descent algorithm. We revisit the mathematical derivations of the Momentum optimizer and discuss the potential problem for steep walls. Inspired by the physical motion of the mass, we propose Smooth Momentum, a new optimizer that improves the behavior on steep walls. We mathematically analyze the characteristics of the proposed optimizer and prove that Smooth Momentum exhibits improved Lipschitz properties and convergence, which allows stable and faster convergence in gradient descent. We also demonstrate how Smooth Gradient, a component of the proposed optimizer, can be plugged into other optimizers, like Adam. The proposed method offers a regularization effect comparable to batch normalization or weight decay. Experiments demonstrate that our proposed optimizer significantly improves the optimization of transformers, convolutional neural networks, and non-convex functions for various tasks and datasets.-
dc.languageEnglish-
dc.publisherKluwer Academic Publishers-
dc.relation.isPartOfApplied Intelligence-
dc.titleSmooth momentum: improving lipschitzness in gradient descent-
dc.typeArticle-
dc.identifier.doi10.1007/s10489-022-04207-7-
dc.type.rimsART-
dc.identifier.bibliographicCitationApplied Intelligence-
dc.identifier.wosid000871171700003-
dc.citation.titleApplied Intelligence-
dc.contributor.affiliatedAuthorBUM, JUN KIM-
dc.contributor.affiliatedAuthorHYEYEON, CHOI-
dc.contributor.affiliatedAuthorJANG, HYEONAH-
dc.contributor.affiliatedAuthorKIM, SANG WOO-
dc.identifier.scopusid2-s2.0-85140446045-
dc.description.journalClass1-
dc.description.journalClass1-
dc.description.isOpenAccessN-
dc.type.docTypeArticle; Early Access-
dc.subject.keywordAuthorGradient descent-
dc.subject.keywordAuthorMomentum-
dc.subject.keywordAuthorLipschitz-
dc.subject.keywordAuthorNon-convex optimization-
dc.subject.keywordAuthorDeep learning-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher

김상우KIM, SANG WOO
Dept of Electrical Enginrg
Read more

Views & Downloads

Browse