TY - GEN
T1 - Demand Charge Control for Energy-intensive Enterprises based on Deep Reinforcement Learning
AU - Wang, Qianwei
AU - Gao, Feng
AU - Liu, Kun
AU - Ming, Fangzhu
AU - Xu, Zhanbo
AU - Wu, Jiang
N1 - Publisher Copyright:
© 2021 IEEE
PY - 2021
Y1 - 2021
N2 - With the advancement of the dual-carbon goal and energy consumption revolution, demand-side management relying on smart grids has gradually become the focus of related research. Due to the obvious surging and high uncertainty of the power load of energy-intensive enterprises, it is difficult to obtain the best control strategy for demand charge. In this paper, we solve the problem of multi-equipment demand charge control for multi-batch tasks with discontinuous production by controlling the load of controllable equipment to reduce electricity costs. To solve the problems of long-time sequence with time coupled and complex systems with difficult modeling, we established a Markov decision process (MDP) model for real-time demand charge control innovatively. To avoid the curse of dimensionality caused by the increasing state space, we introduced the deep Q learning (DQN) algorithm, which successfully solves MDP problems with large state space. Moreover, we introduced constrained deep Q-learning (CDQN) aiming at a large number of action constraints in the problem, which selects the optimal action from the feasible action zone instead of the whole action space to improve the training efficiency and data utilization. Finally, we conducted experiments on simulation case experiments. Under the basic day-ahead production scheduling plan, real-time demand charge control can reduce costs by 10.4% compared with uncontrolled, indicating that this method has achieved excellent performance in obtaining demand charge control strategies.
AB - With the advancement of the dual-carbon goal and energy consumption revolution, demand-side management relying on smart grids has gradually become the focus of related research. Due to the obvious surging and high uncertainty of the power load of energy-intensive enterprises, it is difficult to obtain the best control strategy for demand charge. In this paper, we solve the problem of multi-equipment demand charge control for multi-batch tasks with discontinuous production by controlling the load of controllable equipment to reduce electricity costs. To solve the problems of long-time sequence with time coupled and complex systems with difficult modeling, we established a Markov decision process (MDP) model for real-time demand charge control innovatively. To avoid the curse of dimensionality caused by the increasing state space, we introduced the deep Q learning (DQN) algorithm, which successfully solves MDP problems with large state space. Moreover, we introduced constrained deep Q-learning (CDQN) aiming at a large number of action constraints in the problem, which selects the optimal action from the feasible action zone instead of the whole action space to improve the training efficiency and data utilization. Finally, we conducted experiments on simulation case experiments. Under the basic day-ahead production scheduling plan, real-time demand charge control can reduce costs by 10.4% compared with uncontrolled, indicating that this method has achieved excellent performance in obtaining demand charge control strategies.
KW - deep reinforcement learning
KW - demand charge control
KW - ultra-short-term scheduling
UR - https://www.scopus.com/pages/publications/85128076165
U2 - 10.1109/CAC53003.2021.9728428
DO - 10.1109/CAC53003.2021.9728428
M3 - 会议稿件
AN - SCOPUS:85128076165
T3 - Proceeding - 2021 China Automation Congress, CAC 2021
SP - 6791
EP - 6796
BT - Proceeding - 2021 China Automation Congress, CAC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 China Automation Congress, CAC 2021
Y2 - 22 October 2021 through 24 October 2021
ER -