Deep Reinforcement Learning for Airplane Components Failure Prognostics Full Cycle Automation

##plugins.themes.bootstrap3.article.main##

##plugins.themes.bootstrap3.article.sidebar##

Published Oct 26, 2025
Baoqian Wang Changzhou Wang Denis Osipychev

Abstract

As airplane components degrade over time, airplane service organizations (e.g., Boeing Global Services) and their airline customers need to collaborate on airplane components failure prognostics and replace/maintain components proactively to improve operation efficiency and reduce cost. In particular, airplane service organizations analyze various sensor data that captures the operational states of airplane components to predict possible component failures. Upon identifying an impending component failure, the service organization promptly sends alerts to the airline maintenance team. In response, the airline maintenance team conducts inspections and maintenance on the component and replaces it if necessary.

In this airplane components failure prognostics procedure, machine learning or engineering-based models can be used to make predictions of components failure on each flight. However, it is crucial for airplane service organizations to determine when to send alerts to airlines given the predictions of the full history of flights. Late alerts may cause schedule interruptions or even grounding of the airplane waiting for parts. Early alerts can bring unnecessary inspections that lead to significant cost to airlines. Current solutions rely on heuristics and/or manual engineering reviews to make decisions on sending alerts, which requires significant manual efforts and is difficult to scale.

To improve efficiency of airplane components failure prognostics, we applied deep reinforcement learning (RL) to automate the prognostics procedure while enhancing accuracy of alerts timing. Specifically, we used Long Short-Term Memory (LSTM) neural network model to represent alert policy that outputs alerts decisions based on flight sensor data and interaction history with airlines. To train the alert policy, we built a prognostics environment by using probability models to simulate airplane component state transitions over time and the airline’s feedback to alerts. With this environment, the parameters of alert policy are updated to minimize costs for airlines during the simulated prognostics procedure. This is achieved through the Deep Q-Network algorithm with memory prioritization to mitigate reward sparsity issue. Once learned, the alert policy is deployed to make decisions on sending alerts automatically by consuming incoming flight records and parsing current interactions with airlines. Moreover, we can fine-tune alert policy parameters to incorporate new airplane component features and airline operation changes. We conducted a case study on Boeing 787 air cycle machine (ACM) prognostics, which demonstrated the feasibility and effectiveness of our approach.

How to Cite

Wang, B., Wang, C., & Osipychev, D. (2025). Deep Reinforcement Learning for Airplane Components Failure Prognostics Full Cycle Automation. Annual Conference of the PHM Society, 17(1). https://doi.org/10.36001/phmconf.2025.v17i1.4397
Abstract 0 | PDF Downloads 0

##plugins.themes.bootstrap3.article.details##

Keywords

Airplane Components Failure Prognostics, Deep Reinforcement Learning, Full Cycle Automation

References
Kulkarni, C., Schumann, J., & Roychoudhury, I. (2018). On-board battery monitoring and prognostics for electric-propulsion aircraft. 2018 AIAA/IEEE Electric Aircraft Technologies Symposium (EATS) (pp. 1-12), July 9-11, Cincinnati, Ohio. doi: 10.2514/6.2018-5034
Pidaparthi, B., Jacobs, R., Ghosh, S., Ravi, S. K., Amer, A. W., Luan, L., ... & Wang, L. (2024). Proactive Aircraft Engine Removal Planning with Dynamic Bayesian Networks. Annual Conference of the PHM Society (Vol. 16, No. 1), November 9-14, Nashville, Tennessee. doi: https://doi.org/10.36001/phmconf.2024.v16i1.4148
Dangut, M. D., Skaf, Z., & Jennions, I. K. (2021). An integrated machine learning model for aircraft components rare failure prognostics with log-based dataset. ISA transactions, 113, 127-139. doi:10.1016/j.isatra.2020.05.001
Hu, Y., Miao, X., Zhang, J., Liu, J., & Pan, E. (2021). Reinforcement learning-driven maintenance strategy: A novel solution for long-term aircraft maintenance decision optimization. Computers & industrial engineering, 153. doi:10.1016/j.cje.2020.107056
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., & Hassabis, D. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529-533. doi: https://doi.org/10.1038/nature14236
Van Hasselt, H., Guez, A., & Silver, D. (2016). Deep reinforcement learning with double q-learning. AAAI conference on artificial intelligence (Vol. 30, No. 1), February 12-17, Phoenix, Arizona. doi: https://dl.acm.org/doi/10.5555/3016100.3016191
Schaul, T., Quan, J., Antonoglou, I., & Silver, D. (2015). Prioritized experience replay. arXiv preprint:1511.05952.
D'Agostino, R., & Stephens, M. (2017). Goodness-of-fit-techniques. Routledge.
Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction. MIT Press.
Nair, V., & Hinton, G. E. (2010). Rectified Linear Units Improve Restricted Boltzmann Machines. 27th International Conference on Machine Learning (ICML 2010), June 21-14, Haifa, Israel. doi: https://dl.acm.org/doi/10.5555/3104322.3104425
Graves, A., Fernández, S., Gomez, F., & Schmidhuber, J. (2006). “Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks.” 23rd International Conference on Machine Learning, July 9-15, Pittsburgh, PA. doi: https://dl.acm.org/doi/10.1145/1143844.1143891
Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver. D., & Wierstra, D. (2015). Continuous control with deep reinforcement learning. arXiv preprint:1509.02971.
Section
Industry Experience Papers