TY - GEN
T1 - Deep Reinforcement Learning Based Preventive Maintenance for Wind Turbines
AU - Dong, Wenkang
AU - Zhao, Tianyang
AU - Wu, Yuxin
N1 - Publisher Copyright:
© 2021 IEEE
PY - 2021
Y1 - 2021
N2 - Wind turbines have been the core devices in the low carbon society, where the maintenance cost occupies the critical part of the life cycle cost. A preventive maintenance (PM) problem is formulated to reduce the maintenance cost caused by random failures of WTs, while guaranteeing the reliability of WTs with multiple components, e.g., blades, generators, main bearing, and gearboxes. A linear state transition function is introduced to capture the relation between the PM actions, degradation, and equivalent age of each component. The Weibull distributions and reliability functions are adopted to quantify the reliability level of each component. The state transition and reliability level functions are embedded as the environment function of a deep reinforcement learning (DRL) problem, where the maintenance costs caused by PM actions are treated as the reward function. The DRL problem is further formulated as a Q-network (DQN) with two hidden layers. A practical example is used to verify the effectiveness of the proposed strategy. The results indicate that the learned policy can take advantage of an opportunistic window to maintenance components whose condition is closed to be maintained, reducing the downtime duration and the maintenance costs.
AB - Wind turbines have been the core devices in the low carbon society, where the maintenance cost occupies the critical part of the life cycle cost. A preventive maintenance (PM) problem is formulated to reduce the maintenance cost caused by random failures of WTs, while guaranteeing the reliability of WTs with multiple components, e.g., blades, generators, main bearing, and gearboxes. A linear state transition function is introduced to capture the relation between the PM actions, degradation, and equivalent age of each component. The Weibull distributions and reliability functions are adopted to quantify the reliability level of each component. The state transition and reliability level functions are embedded as the environment function of a deep reinforcement learning (DRL) problem, where the maintenance costs caused by PM actions are treated as the reward function. The DRL problem is further formulated as a Q-network (DQN) with two hidden layers. A practical example is used to verify the effectiveness of the proposed strategy. The results indicate that the learned policy can take advantage of an opportunistic window to maintenance components whose condition is closed to be maintained, reducing the downtime duration and the maintenance costs.
KW - Deep reinforcement learning
KW - Opportunistic maintenance
KW - Preventive maintenance
KW - Wind turbines
UR - https://www.scopus.com/pages/publications/85128187128
U2 - 10.1109/EI252483.2021.9713457
DO - 10.1109/EI252483.2021.9713457
M3 - 会议稿件
AN - SCOPUS:85128187128
T3 - 5th IEEE Conference on Energy Internet and Energy System Integration: Energy Internet for Carbon Neutrality, EI2 2021
SP - 2860
EP - 2865
BT - 5th IEEE Conference on Energy Internet and Energy System Integration
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th IEEE Conference on Energy Internet and Energy System Integration, EI2 2021
Y2 - 22 October 2021 through 25 October 2021
ER -