Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Utilizing Hidden Observations to Enhance the Performance of the Trained Agent

Full metadata record
DC Field Value Language
dc.contributor.authorJang, Sooyoung-
dc.contributor.authorLee, Joohyung-
dc.date.accessioned2022-09-27T06:40:32Z-
dc.date.available2022-09-27T06:40:32Z-
dc.date.created2022-09-22-
dc.date.issued2022-07-
dc.identifier.issn2377-3766-
dc.identifier.urihttps://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/85575-
dc.description.abstractThe frame-skipping strategy has been widely employed in deep reinforcement learning (DRL) technology to train an agent. Specifically, this strategy repeats the action determined by the agent for a fixed number of frames. It increases computational efficiency by reducing the number of inferences by making the action decision sparse. However, previously, these consecutive changes in frames during the frame-skipping were hidden and ignored from the environment and did not affect the agent's action decision. As a result, it can adversely affect the performance of trained agents, where the performance is more critical than computational efficiency. To alleviate these issues, we propose a new framework that utilizes these hidden frames during the frame-skipping, called hidden observation, to enhance the performance of the trained agent. The proposed framework retrieves all hidden observations during frame skipping. It then combines batch inference and an exponentially weighted sum to calculate and merge the outputs from hidden observations. Through experiments, we validated the effectiveness of the proposed method in terms of both performance and stability with only a marginal increase in computation.-
dc.language영어-
dc.language.isoen-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.relation.isPartOfIEEE ROBOTICS AND AUTOMATION LETTERS-
dc.titleUtilizing Hidden Observations to Enhance the Performance of the Trained Agent-
dc.typeArticle-
dc.type.rimsART-
dc.description.journalClass1-
dc.identifier.wosid000838441200010-
dc.identifier.doi10.1109/LRA.2022.3186508-
dc.identifier.bibliographicCitationIEEE ROBOTICS AND AUTOMATION LETTERS, v.7, no.3, pp.7858 - 7864-
dc.description.isOpenAccessN-
dc.identifier.scopusid2-s2.0-85133765546-
dc.citation.endPage7864-
dc.citation.startPage7858-
dc.citation.titleIEEE ROBOTICS AND AUTOMATION LETTERS-
dc.citation.volume7-
dc.citation.number3-
dc.contributor.affiliatedAuthorLee, Joohyung-
dc.type.docTypeArticle-
dc.subject.keywordAuthorAI-Enabled Robotics-
dc.subject.keywordAuthorRobust/Adaptive Control-
dc.subject.keywordAuthorControl Architectures and Programming-
dc.subject.keywordAuthorReinforcement Learning-
dc.relation.journalResearchAreaRobotics-
dc.relation.journalWebOfScienceCategoryRobotics-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
Files in This Item
There are no files associated with this item.
Appears in
Collections
IT융합대학 > 소프트웨어학과 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Joo Hyung photo

Lee, Joo Hyung
College of IT Convergence (Department of Software)
Read more

Altmetrics

Total Views & Downloads

BROWSE