Explainable Artificial Intelligence for Fault Diagnosis of Industrial Processes

Kyojin Jang, Karl Ezra Salgado Pilario, Nayoung Lee, Il Moon, Jonggeol Na

Research output: Contribution to journalArticlepeer-review

17 Scopus citations

Abstract

Process monitoring is important for ensuring operational reliability and preventing occupational accidents. In recent years, data-driven methods such as machine learning and deep learning have been preferred for fault detection and diagnosis. In particular, unsupervised learning algorithms, such as auto-encoders, exhibit good detection performance, even for unlabeled data from complex processes. However, decisions generated from deep-neural-network-based models are difficult to interpret and cannot provide explanatory insight to users. We address this issue by proposing a new fault diagnosis method using explainable artificial intelligence to break the traditional trade-off between the accuracy and interpretability of deep learning model. First, an adversarial auto-encoder model for fault detection is built and then interpreted through the integration of Shapley additive explanations (SHAP) with a combined monitoring index. Using SHAP values, a diagnosis is conducted by allocating credit for detected faults, deviations from a normal state, among its input variables. The proposed diagnosis method can consider not only reconstruction space but also latent space unlike conventional method, which evaluate only reconstruction error. The proposed method was applied to two chemical process systems and compared with conventional diagnosis methods. The results highlight that the proposed method achieves the exact fault diagnosis for single and multiple faults and, also, distinguishes the global pattern of various fault types.

Original languageEnglish
Pages (from-to)1-8
Number of pages8
JournalIEEE Transactions on Industrial Informatics
DOIs
StateAccepted/In press - 2023

Bibliographical note

Publisher Copyright:
IEEE

Keywords

  • CSTR
  • Computational modeling
  • Data models
  • Explainable AI
  • Fault detection
  • Fault diagnosis
  • Indexes
  • Modeling
  • Monitoring
  • Tennessee Eastman process
  • auto-encoder
  • fault diagnosis

Fingerprint

Dive into the research topics of 'Explainable Artificial Intelligence for Fault Diagnosis of Industrial Processes'. Together they form a unique fingerprint.

Cite this