Open Access System for Information Sharing

Login Library

 

Thesis
Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.author권순철-
dc.date.accessioned2023-08-31T16:30:43Z-
dc.date.available2023-08-31T16:30:43Z-
dc.date.issued2023-
dc.identifier.otherOAK-2015-10029-
dc.identifier.urihttp://postech.dcollection.net/common/orgView/200000660074ko_KR
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/118226-
dc.descriptionDoctor-
dc.description.abstractGrammatical error correction (GEC) has been successful with deep and complex neural machine translation models, but published annotated datasets to train the large models are scarce. In this dissertation, I propose a novel self-feeding training method that generates incorrect sentences from correct sentences. The proposed training method can generate appropriate wrong sentences from unlabeled sentences, using a data generation model trained as an autoencoder. It can also add artificial noise to correct sentences to automatically generate noisy sentences. I show that the GEC models trained with the self-feeding training method are successful without extra annotated data or deeper neural network-based models, achieving F0.5 score of 0.5982 on the CoNLL-2014 Shared Task test data with a transformer model. The results also show that fully unlabeled training is possible for data-scarce domains and languages.-
dc.languageeng-
dc.publisher포항공과대학교-
dc.titleSelf-feeding Semi-supervised Training Method for Grammatical Error Correction-
dc.typeThesis-
dc.contributor.college컴퓨터공학과-
dc.date.degree2023- 2-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Views & Downloads

Browse