Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

An Enhanced Proximal Policy Optimization-Based Reinforcement Learning Method with Random Forest for Hyperparameter Optimization

Full metadata record
DC Field Value Language
dc.contributor.authorMa, Zhixin-
dc.contributor.authorCui, Shengmin-
dc.contributor.authorJoe, Inwhee-
dc.date.accessioned2023-07-05T03:53:22Z-
dc.date.available2023-07-05T03:53:22Z-
dc.date.created2022-09-08-
dc.date.issued2022-07-
dc.identifier.issn2076-3417-
dc.identifier.urihttps://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/186187-
dc.description.abstractFor most machine learning and deep learning models, the selection of hyperparameters has a significant impact on the performance of the model. Therefore, deep learning and data analysis experts have to spend a lot of time on hyperparameter tuning when building a model for accomplishing a task. Although there are many algorithms used to solve hyperparameter optimization (HPO), these methods require the results of the actual trials at each epoch to help perform the search. To reduce the number of trials, model-based reinforcement learning adopts multilayer perceptron (MLP) to capture the relationship between hyperparameter settings and model performance. However, MLP needs to be carefully designed because there is a risk of overfitting. Thus, we propose a random forest-enhanced proximal policy optimization (RFEPPO) reinforcement learning algorithm to solve the HPO problem. In addition, reinforcement learning as a solution to HPO will encounter the sparse reward problem, eventually leading to slow convergence. To address this problem, we employ the intrinsic reward, which introduces the prediction error as the reward signal. Experiments carried on nine tabular datasets and two image classification datasets demonstrate the effectiveness of our model.-
dc.language영어-
dc.language.isoen-
dc.publisherMDPI-
dc.titleAn Enhanced Proximal Policy Optimization-Based Reinforcement Learning Method with Random Forest for Hyperparameter Optimization-
dc.typeArticle-
dc.contributor.affiliatedAuthorJoe, Inwhee-
dc.identifier.doi10.3390/app12147006-
dc.identifier.scopusid2-s2.0-85137367350-
dc.identifier.wosid000834406900001-
dc.identifier.bibliographicCitationAPPLIED SCIENCES-BASEL, v.12, no.14, pp.1 - 20-
dc.relation.isPartOfAPPLIED SCIENCES-BASEL-
dc.citation.titleAPPLIED SCIENCES-BASEL-
dc.citation.volume12-
dc.citation.number14-
dc.citation.startPage1-
dc.citation.endPage20-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaChemistry-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaMaterials Science-
dc.relation.journalResearchAreaPhysics-
dc.relation.journalWebOfScienceCategoryChemistry, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryEngineering, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryMaterials Science, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryPhysics, Applied-
dc.subject.keywordAuthorhyperparameter optimization (HPO)-
dc.subject.keywordAuthorproximal policy optimization-
dc.subject.keywordAuthorrandom forest-
dc.subject.keywordAuthorreinforcement learning-
dc.identifier.urlhttps://www.mdpi.com/2076-3417/12/14/7006-
Files in This Item
Appears in
Collections
서울 공과대학 > 서울 컴퓨터소프트웨어학부 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Joe, Inwhee photo

Joe, Inwhee
COLLEGE OF ENGINEERING (SCHOOL OF COMPUTER SCIENCE)
Read more

Altmetrics

Total Views & Downloads

BROWSE