DC Field | Value | Language |
---|---|---|
dc.contributor.author | PARK, JONG HYEOK | - |
dc.contributor.author | JEON, SOO | - |
dc.contributor.author | HAN, SOOHEE | - |
dc.date.accessioned | 2024-01-18T08:50:04Z | - |
dc.date.available | 2024-01-18T08:50:04Z | - |
dc.date.created | 2023-12-01 | - |
dc.date.issued | 2023-11 | - |
dc.identifier.issn | 0278-0046 | - |
dc.identifier.uri | https://oasis.postech.ac.kr/handle/2014.oak/119762 | - |
dc.description.abstract | This article proposes a data-efficient model-based reinforcement learning (RL) algorithm empowered by reliable future reward estimates achieved through a confidence-based probabilistic ensemble terminal critics (PETC). The proposed algorithm utilizes a model-predictive controller to choose an action that optimizes the sum of the near and distant future rewards for a given current state. Near future rewards with high confidence are determined directly from trained deterministic dynamics and reward models. Distant future rewards beyond these horizons are meticulously assessed using the proposed confidence-based PETC, which minimizes estimation errors inherent in the distant future and quantifies uncertainty confidence. Through such confidence-based guided actions, the proposed approach is expected to operate in a reliable, explainable, and data-efficient manner, consistently guiding the system to an optimal trajectory. A comparison with the existing state-of-the-art RL algorithms for eight DeepMind Control Suite tasks confirms the superior data efficiency of the proposed approach, which achieves an average cumulative reward of 761.2 in merely 500K steps, whereas the other algorithms score below 700.0. The proposed algorithm is also successfully applied to two real-world control applications, namely single- and double-cartpole swing-up tasks. | - |
dc.language | English | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.relation.isPartOf | IEEE Transactions on Industrial Electronics | - |
dc.title | Model-Based Reinforcement Learning With Probabilistic Ensemble Terminal Critics for Data-Efficient Control Applications | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TIE.2023.3331074 | - |
dc.type.rims | ART | - |
dc.identifier.bibliographicCitation | IEEE Transactions on Industrial Electronics, v.71, no.8, pp.9470 - 9479 | - |
dc.identifier.wosid | 001122808200001 | - |
dc.citation.endPage | 9479 | - |
dc.citation.number | 8 | - |
dc.citation.startPage | 9470 | - |
dc.citation.title | IEEE Transactions on Industrial Electronics | - |
dc.citation.volume | 71 | - |
dc.contributor.affiliatedAuthor | PARK, JONG HYEOK | - |
dc.contributor.affiliatedAuthor | HAN, SOOHEE | - |
dc.identifier.scopusid | 2-s2.0-85178067053 | - |
dc.description.journalClass | 1 | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | N | - |
dc.type.docType | Article; Early Access | - |
dc.subject.keywordAuthor | Heuristic algorithms | - |
dc.subject.keywordAuthor | Data models | - |
dc.subject.keywordAuthor | Robots | - |
dc.subject.keywordAuthor | Probabilistic logic | - |
dc.subject.keywordAuthor | Computational modeling | - |
dc.subject.keywordAuthor | Reliability | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Cartpole system | - |
dc.subject.keywordAuthor | model-predictive controller (MPC) | - |
dc.subject.keywordAuthor | model-based reinforcement learning (RL) | - |
dc.subject.keywordAuthor | probabilistic ensemble terminal critics (PETC) | - |
dc.relation.journalWebOfScienceCategory | Automation & Control Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Instruments & Instrumentation | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Automation & Control Systems | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Instruments & Instrumentation | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.
library@postech.ac.kr Tel: 054-279-2548
Copyrights © by 2017 Pohang University of Science ad Technology All right reserved.