Deep Reinforcement Learning for Airplane Components Failure Prognostics Full Cycle Automation
##plugins.themes.bootstrap3.article.main##
##plugins.themes.bootstrap3.article.sidebar##
Abstract
As airplane components degrade over time, airplane service organizations (e.g., Boeing Global Services) and their airline customers need to collaborate on airplane components failure prognostics and replace/maintain components proactively to improve operation efficiency and reduce cost. In particular, airplane service organizations analyze various sensor data that captures the operational states of airplane components to predict possible component failures. Upon identifying an impending component failure, the service organization promptly sends alerts to the airline maintenance team. In response, the airline maintenance team conducts inspections and maintenance on the component and replaces it if necessary.
In this airplane components failure prognostics procedure, machine learning or engineering-based models can be used to make predictions of components failure on each flight. However, it is crucial for airplane service organizations to determine when to send alerts to airlines given the predictions of the full history of flights. Late alerts may cause schedule interruptions or even grounding of the airplane waiting for parts. Early alerts can bring unnecessary inspections that lead to significant cost to airlines. Current solutions rely on heuristics and/or manual engineering reviews to make decisions on sending alerts, which requires significant manual efforts and is difficult to scale.
To improve efficiency of airplane components failure prognostics, we applied deep reinforcement learning (RL) to automate the prognostics procedure while enhancing accuracy of alerts timing. Specifically, we used Long Short-Term Memory (LSTM) neural network model to represent alert policy that outputs alerts decisions based on flight sensor data and interaction history with airlines. To train the alert policy, we built a prognostics environment by using probability models to simulate airplane component state transitions over time and the airline’s feedback to alerts. With this environment, the parameters of alert policy are updated to minimize costs for airlines during the simulated prognostics procedure. This is achieved through the Deep Q-Network algorithm with memory prioritization to mitigate reward sparsity issue. Once learned, the alert policy is deployed to make decisions on sending alerts automatically by consuming incoming flight records and parsing current interactions with airlines. Moreover, we can fine-tune alert policy parameters to incorporate new airplane component features and airline operation changes. We conducted a case study on Boeing 787 air cycle machine (ACM) prognostics, which demonstrated the feasibility and effectiveness of our approach.
How to Cite
##plugins.themes.bootstrap3.article.details##
Airplane Components Failure Prognostics, Deep Reinforcement Learning, Full Cycle Automation
Pidaparthi, B., Jacobs, R., Ghosh, S., Ravi, S. K., Amer, A. W., Luan, L., ... & Wang, L. (2024). Proactive Aircraft Engine Removal Planning with Dynamic Bayesian Networks. Annual Conference of the PHM Society (Vol. 16, No. 1), November 9-14, Nashville, Tennessee. doi: https://doi.org/10.36001/phmconf.2024.v16i1.4148
Dangut, M. D., Skaf, Z., & Jennions, I. K. (2021). An integrated machine learning model for aircraft components rare failure prognostics with log-based dataset. ISA transactions, 113, 127-139. doi:10.1016/j.isatra.2020.05.001
Hu, Y., Miao, X., Zhang, J., Liu, J., & Pan, E. (2021). Reinforcement learning-driven maintenance strategy: A novel solution for long-term aircraft maintenance decision optimization. Computers & industrial engineering, 153. doi:10.1016/j.cje.2020.107056
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., & Hassabis, D. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529-533. doi: https://doi.org/10.1038/nature14236
Van Hasselt, H., Guez, A., & Silver, D. (2016). Deep reinforcement learning with double q-learning. AAAI conference on artificial intelligence (Vol. 30, No. 1), February 12-17, Phoenix, Arizona. doi: https://dl.acm.org/doi/10.5555/3016100.3016191
Schaul, T., Quan, J., Antonoglou, I., & Silver, D. (2015). Prioritized experience replay. arXiv preprint:1511.05952.
D'Agostino, R., & Stephens, M. (2017). Goodness-of-fit-techniques. Routledge.
Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction. MIT Press.
Nair, V., & Hinton, G. E. (2010). Rectified Linear Units Improve Restricted Boltzmann Machines. 27th International Conference on Machine Learning (ICML 2010), June 21-14, Haifa, Israel. doi: https://dl.acm.org/doi/10.5555/3104322.3104425
Graves, A., Fernández, S., Gomez, F., & Schmidhuber, J. (2006). “Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks.” 23rd International Conference on Machine Learning, July 9-15, Pittsburgh, PA. doi: https://dl.acm.org/doi/10.1145/1143844.1143891
Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver. D., & Wierstra, D. (2015). Continuous control with deep reinforcement learning. arXiv preprint:1509.02971.

This work is licensed under a Creative Commons Attribution 3.0 Unported License.
The Prognostic and Health Management Society advocates open-access to scientific data and uses a Creative Commons license for publishing and distributing any papers. A Creative Commons license does not relinquish the author’s copyright; rather it allows them to share some of their rights with any member of the public under certain conditions whilst enjoying full legal protection. By submitting an article to the International Conference of the Prognostics and Health Management Society, the authors agree to be bound by the associated terms and conditions including the following:
As the author, you retain the copyright to your Work. By submitting your Work, you are granting anybody the right to copy, distribute and transmit your Work and to adapt your Work with proper attribution under the terms of the Creative Commons Attribution 3.0 United States license. You assign rights to the Prognostics and Health Management Society to publish and disseminate your Work through electronic and print media if it is accepted for publication. A license note citing the Creative Commons Attribution 3.0 United States License as shown below needs to be placed in the footnote on the first page of the article.
First Author et al. This is an open-access article distributed under the terms of the Creative Commons Attribution 3.0 United States License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.