Open Access System for Information Sharing

Login Library

 

Conference
Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.authorJang, Jinhyeok-
dc.contributor.authorYun, Woo-Han-
dc.contributor.authorKim, Won Hwa-
dc.contributor.authorYoon, Youngwoo-
dc.contributor.authorKim, Jaehong-
dc.contributor.authorLee, Jaeyeon-
dc.contributor.authorHan, ByungOk-
dc.date.accessioned2024-03-06T01:06:14Z-
dc.date.available2024-03-06T01:06:14Z-
dc.date.created2024-02-20-
dc.date.issued2023-07-25-
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/121291-
dc.description.abstractRecent complicated problems require large-scale datasets and complex model architectures, however, it is difficult to train such large networks due to high computational issues. Significant efforts have been made to make the training more efficient such as momentum, learning rate scheduling, weight regularization, and meta-learning. Based on our observations on 1) high correlation between past weights and future weights, 2) conditions for beneficial weight prediction, and 3) feasibility of weight prediction, we propose a more general framework by intermittently skipping a handful of epochs by periodically forecasting near future weights, i.e., a Weight Nowcaster Network (WNN). As an add-on module, WNN predicts the future weights to make the learning process faster regardless of tasks and architectures. Experimental results show that WNN can significantly save actual time cost for training with an additional marginal time to train WNN. We validate the generalization capability of WNN under various tasks, and demonstrate that it works well even for unseen tasks. The code and pre-trained model are available at https://github.com/jjh6297/WNN.-
dc.languageEnglish-
dc.publisherML Research Press-
dc.relation.isPartOfInternational Conference on Machine Learning (ICML)-
dc.relation.isPartOfProceedings of Machine Learning Research-
dc.titleLearning to Boost Training by Periodic Nowcasting Near Future Weights-
dc.typeConference-
dc.type.rimsCONF-
dc.identifier.bibliographicCitationInternational Conference on Machine Learning (ICML), pp.14730 - 14757-
dc.citation.conferenceDate2023-07-23-
dc.citation.conferencePlaceUS-
dc.citation.endPage14757-
dc.citation.startPage14730-
dc.citation.titleInternational Conference on Machine Learning (ICML)-
dc.contributor.affiliatedAuthorKim, Won Hwa-
dc.description.journalClass1-
dc.description.journalClass1-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Views & Downloads

Browse