Deep reinforcement learning for cooperative robots based on adaptive sentiment feedback
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Jeon, Haein | - |
dc.contributor.author | Kim, Dae-Won | - |
dc.contributor.author | Kang, Bo-Yeong | - |
dc.date.accessioned | 2024-01-24T05:01:28Z | - |
dc.date.available | 2024-01-24T05:01:28Z | - |
dc.date.issued | 2024-06 | - |
dc.identifier.issn | 0957-4174 | - |
dc.identifier.issn | 1873-6793 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/71331 | - |
dc.description.abstract | Human–robot cooperative tasks have gained importance with the emergence of robotics and artificial intelligence technology. In interactive reinforcement learning techniques, robots learn target tasks by receiving feedback from an experienced human trainer. However, most interactive reinforcement learning studies require a separate process to integrate the trainer's feedback into the training dataset, making it challenging for robots to learn new tasks from humans in real-time. Furthermore, the types of feedback sentences that trainers can use are limited in previous research. To address these limitations, this paper proposes a robot teaching strategy that uses deep RL via human–robot interaction to learn table balancing tasks interactively. The proposed system employs Deep Q-Network with real-time sentiment feedback delivered through the trainer's speech to learn cooperative tasks. We designed a novel reward function that incorporates sentiment feedback from human speech in real-time during the learning process. The paper presents an improved reward shaping technique based on subdivided feedback levels and shrinking feedback. This function serves as a guide for the robot to engage in natural interactions with humans and enables it to learn the tasks effectively. Experimental results demonstrate that the proposed interactive deep reinforcement learning model achieved a high success rate of up to 99.06%, outperforming the model without sentiment feedback. © 2023 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | Elsevier Ltd | - |
dc.title | Deep reinforcement learning for cooperative robots based on adaptive sentiment feedback | - |
dc.type | Article | - |
dc.identifier.doi | 10.1016/j.eswa.2023.121198 | - |
dc.identifier.bibliographicCitation | Expert Systems with Applications, v.243 | - |
dc.description.isOpenAccess | N | - |
dc.identifier.wosid | 001139775500001 | - |
dc.identifier.scopusid | 2-s2.0-85179581967 | - |
dc.citation.title | Expert Systems with Applications | - |
dc.citation.volume | 243 | - |
dc.type.docType | Article | - |
dc.publisher.location | 영국 | - |
dc.subject.keywordAuthor | Deep reinforcement learning | - |
dc.subject.keywordAuthor | Human-in-the-loop | - |
dc.subject.keywordAuthor | Human–robot interaction | - |
dc.subject.keywordAuthor | Interactive reinforcement learning | - |
dc.subject.keywordAuthor | Reward shaping | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Operations Research & Management Science | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Operations Research & Management Science | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
84, Heukseok-ro, Dongjak-gu, Seoul, Republic of Korea (06974)02-820-6194
COPYRIGHT 2019 Chung-Ang University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.