Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Demand Response Management for Industrial Facilities: A Deep Reinforcement Learning Approach

Authors
Huang, XuefeiHong, Seung HoYu, MengmengDing, YueminJiang, Junhui
Issue Date
Jun-2019
Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Keywords
Artificial intelligence; deep reinforcement learning; demand response (DR); industrial facilities; actor-critic
Citation
IEEE ACCESS, v.7, pp 82194 - 82205
Pages
12
Indexed
SCIE
SCOPUS
Journal Title
IEEE ACCESS
Volume
7
Start Page
82194
End Page
82205
URI
https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/2771
DOI
10.1109/ACCESS.2019.2924030
ISSN
2169-3536
2169-3536
Abstract
As a major consumer of energy, the industrial sector must assume the responsibility for improving energy efficiency and reducing carbon emissions. However, most existing studies on industrial energy management are suffering from modeling complex industrial processes. To address this issue, a model-free demand response (DR) scheme for industrial facilities was developed. In practical terms, we first formulated the Markov decision process (MDP) for industrial DR, which presents the composition of the state, action, and reward function in detail. Then, we designed an actor-critic-based deep reinforcement learning algorithm to determine the optimal energy management policy, where both the actor (Policy) and the critic (Value function) are implemented by the deep neural network. We then confirmed the validity of our scheme by applying it to a real-world industry. Our algorithm identified an optimal energy consumption schedule, reducing energy costs without compromising production.
Files in This Item
Go to Link
Appears in
Collections
COLLEGE OF ENGINEERING SCIENCES > SCHOOL OF ELECTRICAL ENGINEERING > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE