TY - JOUR
T1 - A Remaining Useful Life Prediction Method of Rolling Bearings Based on Deep Reinforcement Learning
AU - Zheng, Guokang
AU - Li, Yasong
AU - Zhou, Zheng
AU - Yan, Ruqiang
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024/7/1
Y1 - 2024/7/1
N2 - Remaining useful life (RUL) prediction technology is a crucial task in prognostics and health management (PHM) systems, as it contributes to the enhancement of the reliability of equipment operation. With the development of Industrial Internet of Things (IIoT) technologies, it becomes possible to efficiently coordinate data collection for mechanical equipment, enabling real-time monitoring of device status and performance. This could provide more accurate estimations of the RUL. Although current RUL prediction techniques predominantly rely on deep learning (DL), these approaches often neglect the temporal correlation within training samples, resulting in unstable prediction outcomes. To address this issue, a novel RUL prediction method is introduced, leveraging deep reinforcement learning (DRL). This method combines the effective feature extraction ability of DL with the preservation of temporal correlation between samples through reinforcement learning. First, an autoencoder (AE) is employed to extract key features that are most relevant to degenerative process from the original signals collected from mechanical equipment. Second, state variables in reinforcement learning are constructed using the extracted features and the predicted RUL value of the sample at the previous time step. Finally, a DRL model based on the twin delayed deep deterministic policy gradient (DDPG) algorithm (TD3) is trained after setting an appropriate action space and reward function. Validation using XJTU-SY bearing data set demonstrates that the DRL method yields lesser root mean square error (RMSE) and more stable prediction results compared to alternative methods.
AB - Remaining useful life (RUL) prediction technology is a crucial task in prognostics and health management (PHM) systems, as it contributes to the enhancement of the reliability of equipment operation. With the development of Industrial Internet of Things (IIoT) technologies, it becomes possible to efficiently coordinate data collection for mechanical equipment, enabling real-time monitoring of device status and performance. This could provide more accurate estimations of the RUL. Although current RUL prediction techniques predominantly rely on deep learning (DL), these approaches often neglect the temporal correlation within training samples, resulting in unstable prediction outcomes. To address this issue, a novel RUL prediction method is introduced, leveraging deep reinforcement learning (DRL). This method combines the effective feature extraction ability of DL with the preservation of temporal correlation between samples through reinforcement learning. First, an autoencoder (AE) is employed to extract key features that are most relevant to degenerative process from the original signals collected from mechanical equipment. Second, state variables in reinforcement learning are constructed using the extracted features and the predicted RUL value of the sample at the previous time step. Finally, a DRL model based on the twin delayed deep deterministic policy gradient (DDPG) algorithm (TD3) is trained after setting an appropriate action space and reward function. Validation using XJTU-SY bearing data set demonstrates that the DRL method yields lesser root mean square error (RMSE) and more stable prediction results compared to alternative methods.
KW - Deep reinforcement learning (DRL)
KW - Industrial Internet of Things (IIoT)
KW - prognostics and health management (PHM)
KW - remaining useful life (RUL)
UR - https://www.scopus.com/pages/publications/85184803400
U2 - 10.1109/JIOT.2024.3363610
DO - 10.1109/JIOT.2024.3363610
M3 - 文章
AN - SCOPUS:85184803400
SN - 2327-4662
VL - 11
SP - 22938
EP - 22949
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 13
ER -