Open Access System for Information Sharing

Login Library

 

Conference
Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.authorKwon, Eunji-
dc.contributor.authorSong, Haena-
dc.contributor.authorPark, Jihye-
dc.contributor.authorKang, Seokhyeong-
dc.date.accessioned2024-03-06T08:10:43Z-
dc.date.available2024-03-06T08:10:43Z-
dc.date.created2024-02-22-
dc.date.issued2023-04-17-
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/122623-
dc.description.abstractIt is difficult to employ transformer models for computer vision in mobile devices due to their memory- and computation-intensive properties. Accordingly, there is ongoing research on various methods for compressing transformer models, such as pruning. However, general computing platforms such as central processing units (CPUs) and graphics processing units (GPUs) are not energy-efficient to accelerate the pruned model due to their structured sparsity. This paper proposes a low-power accelerator for transformers with various sizes of structured sparsity induced by pruning with different granularity. In this study, we can accelerate a transformer that has been pruned in a head-wise, line-wise, or block-wise manner. We developed a head scheduling algorithm to support head-wise skip operations and resolve the processing engine (PE) load imbalance problem caused by different number of operations in one head. Moreover, we implemented a sparse general matrix-to-matrix multiplication (sparse GEMM) module that supports line-wise and block-wise skipping. As a result, when compared with a mobile GPU and mobile CPU respectively, our proposed accelerator achieved 6.1× and 13.6× improvements in energy efficiency for the detection transformer (DETR) model and achieved approximately 2.6× and 7.9× improvements in the energy efficiency on average for the vision transformer (ViT) models.-
dc.languageEnglish-
dc.publisherInstitute of Electrical and Electronics Engineers Inc.-
dc.relation.isPartOf2023 Design, Automation and Test in Europe Conference and Exhibition, DATE 2023-
dc.relation.isPartOfProceedings -Design, Automation and Test in Europe, DATE-
dc.titleMobile Accelerator Exploiting Sparsity of Multi-Heads, Lines, and Blocks in Transformers in Computer Vision-
dc.typeConference-
dc.type.rimsCONF-
dc.identifier.bibliographicCitation2023 Design, Automation and Test in Europe Conference and Exhibition, DATE 2023-
dc.citation.conferenceDate2023-04-17-
dc.citation.conferencePlaceBE-
dc.citation.title2023 Design, Automation and Test in Europe Conference and Exhibition, DATE 2023-
dc.contributor.affiliatedAuthorKang, Seokhyeong-
dc.description.journalClass1-
dc.description.journalClass1-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Views & Downloads

Browse