Abstract
Process monitoring is important for ensuring operational reliability and preventing occupational accidents. In recent years, data-driven methods such as machine learning and deep learning have been preferred for fault detection and diagnosis. In particular, unsupervised learning algorithms, such as auto-encoders, exhibit good detection performance, even for unlabeled data from complex processes. However, decisions generated from deep-neural-network-based models are difficult to interpret and cannot provide explanatory insight to users. We address this issue by proposing a new fault diagnosis method using explainable artificial intelligence to break the traditional trade-off between the accuracy and interpretability of deep learning model. First, an adversarial auto-encoder model for fault detection is built and then interpreted through the integration of Shapley additive explanations (SHAP) with a combined monitoring index. Using SHAP values, a diagnosis is conducted by allocating credit for detected faults, deviations from a normal state, among its input variables. The proposed diagnosis method can consider not only reconstruction space but also latent space unlike conventional method, which evaluate only reconstruction error. The proposed method was applied to two chemical process systems and compared with conventional diagnosis methods. The results highlight that the proposed method achieves the exact fault diagnosis for single and multiple faults and, also, distinguishes the global pattern of various fault types.
Original language | English |
---|---|
Pages (from-to) | 1-8 |
Number of pages | 8 |
Journal | IEEE Transactions on Industrial Informatics |
DOIs | |
State | Accepted/In press - 2023 |
Bibliographical note
Publisher Copyright:IEEE
Keywords
- CSTR
- Computational modeling
- Data models
- Explainable AI
- Fault detection
- Fault diagnosis
- Indexes
- Modeling
- Monitoring
- Tennessee Eastman process
- auto-encoder
- fault diagnosis