A Case-study Led Investigation of Explainable AI (XAI) to Support Deployment of Prognostics in the industry

##plugins.themes.bootstrap3.article.main##

##plugins.themes.bootstrap3.article.sidebar##

Published Jun 29, 2022
Omnia Amin Blair Brown Bruce Stephen Stephen McArthur

Abstract

Civil nuclear generation plant must maximise it’s operational uptime in order to maintain it’s viability. With aging plant and heavily regulated operating constraints, monitoring is commonplace, but identifying health indicators to pre-empt disruptive faults is challenging owing to the volumes of data involved. Machine learning (ML) models are increasingly deployed in prognostics and health management (PHM) systems in various industrial applications, however, many of these are black box models that provide good performance but little or no insight into how predictions are reached. In nuclear generation, there is significant regulatory oversight and therefore a necessity to explain decisions based on outputs from predictive models. These explanations can then enable stakeholders to trust these outputs, satisfy regulatory bodies and subsequently make more effective operational decisions. How ML model outputs convey explanations to stakeholders is important, so these explanations must be in human (and technical domain related) understandable terms. Consequently, stakeholders can rapidly interpret, then trust predictions better, and will be able to act on them more effectively. The main contributions of this paper are: 1. introduce XAI into the PHM of industrial assets and provide a novel set of algorithms that translate the explanations produced by SHAP to text-based human-interpretable explanations; and 2. consider the context of these explanations as intended for application to prognostics of critical assets in industrial applications. The use of XAI will not only help in understanding how these ML models work, but also describe the most important features contributing to predicted degradation of the nuclear generation asset.

How to Cite

Amin, O. ., Brown, B. ., Stephen, B. ., & McArthur, S. . (2022). A Case-study Led Investigation of Explainable AI (XAI) to Support Deployment of Prognostics in the industry. PHM Society European Conference, 7(1), 9–20. https://doi.org/10.36001/phme.2022.v7i1.3336
Abstract 544 | PDF Downloads 456

##plugins.themes.bootstrap3.article.details##

Keywords

Prognostics, Machine learning, Explainable AI, assets

Section
Technical Papers