Open Access System for Information Sharing

Login Library

 

Article
Cited 2 time in webofscience Cited 3 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.authorPARK, JONG HYEOK-
dc.contributor.authorJEON, SOO-
dc.contributor.authorHAN, SOOHEE-
dc.date.accessioned2024-01-18T08:50:04Z-
dc.date.available2024-01-18T08:50:04Z-
dc.date.created2023-12-01-
dc.date.issued2023-11-
dc.identifier.issn0278-0046-
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/119762-
dc.description.abstractThis article proposes a data-efficient model-based reinforcement learning (RL) algorithm empowered by reliable future reward estimates achieved through a confidence-based probabilistic ensemble terminal critics (PETC). The proposed algorithm utilizes a model-predictive controller to choose an action that optimizes the sum of the near and distant future rewards for a given current state. Near future rewards with high confidence are determined directly from trained deterministic dynamics and reward models. Distant future rewards beyond these horizons are meticulously assessed using the proposed confidence-based PETC, which minimizes estimation errors inherent in the distant future and quantifies uncertainty confidence. Through such confidence-based guided actions, the proposed approach is expected to operate in a reliable, explainable, and data-efficient manner, consistently guiding the system to an optimal trajectory. A comparison with the existing state-of-the-art RL algorithms for eight DeepMind Control Suite tasks confirms the superior data efficiency of the proposed approach, which achieves an average cumulative reward of 761.2 in merely 500K steps, whereas the other algorithms score below 700.0. The proposed algorithm is also successfully applied to two real-world control applications, namely single- and double-cartpole swing-up tasks.-
dc.languageEnglish-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.isPartOfIEEE Transactions on Industrial Electronics-
dc.titleModel-Based Reinforcement Learning With Probabilistic Ensemble Terminal Critics for Data-Efficient Control Applications-
dc.typeArticle-
dc.identifier.doi10.1109/TIE.2023.3331074-
dc.type.rimsART-
dc.identifier.bibliographicCitationIEEE Transactions on Industrial Electronics, v.71, no.8, pp.9470 - 9479-
dc.identifier.wosid001122808200001-
dc.citation.endPage9479-
dc.citation.number8-
dc.citation.startPage9470-
dc.citation.titleIEEE Transactions on Industrial Electronics-
dc.citation.volume71-
dc.contributor.affiliatedAuthorPARK, JONG HYEOK-
dc.contributor.affiliatedAuthorHAN, SOOHEE-
dc.identifier.scopusid2-s2.0-85178067053-
dc.description.journalClass1-
dc.description.journalClass1-
dc.description.isOpenAccessN-
dc.type.docTypeArticle; Early Access-
dc.subject.keywordAuthorHeuristic algorithms-
dc.subject.keywordAuthorData models-
dc.subject.keywordAuthorRobots-
dc.subject.keywordAuthorProbabilistic logic-
dc.subject.keywordAuthorComputational modeling-
dc.subject.keywordAuthorReliability-
dc.subject.keywordAuthorReinforcement learning-
dc.subject.keywordAuthorCartpole system-
dc.subject.keywordAuthormodel-predictive controller (MPC)-
dc.subject.keywordAuthormodel-based reinforcement learning (RL)-
dc.subject.keywordAuthorprobabilistic ensemble terminal critics (PETC)-
dc.relation.journalWebOfScienceCategoryAutomation & Control Systems-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryInstruments & Instrumentation-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaAutomation & Control Systems-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaInstruments & Instrumentation-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher

한수희HAN, SOOHEE
Dept of Electrical Enginrg
Read more

Views & Downloads

Browse