Acessibilidade / Reportar erro

An approach to robust condition monitoring in industrial processes using pythagorean membership grades

Abstract

In this paper, a robust approach to improve the performance of a condition monitoring process in industrial plants by using Pythagorean membership grades is presented. The FCM algorithm is modified by using Pythagorean fuzzy sets, to obtain a new variant of it called Pythagorean Fuzzy C-Means (PyFCM). In addition, a kernel version of PyFCM (KPyFCM) is obtained in order to achieve greater separability among classes, and reduce classification errors. The approach proposed is validated using experimental datasets and the Tennessee Eastman (TE) process benchmark. The results are compared with the results obtained with other algorithms that use standard and non-standard membership grades. The highest performance obtained by the approach proposed indicate its feasibility.

Key words
Robust diagnostic approach; industrial plants; fuzzy algorithms; pythagorean fuzzy sets

INTRODUCTION

Nowadays, there is a marked necessity in industrial plants to produce with higher quality in order to comply with environmental and industrial regulations (Hwang et al. 2010HWANG I, KIM S, KIM Y & SEAH C. 2010. A Survey of Fault Detection, Isolation, and Reconfiguration Methods. IEEE Control Syst 18: 636-656., Venkatasubramanian et al. 2003a). However, the faults in equipments can have an unfavorable impact on the availability of systems, environment and the safety of operators. For such reasons, the faults need to be detected and isolated using an effective condition monitoring system (Isermann 2011ISERMANN R. 2011. Fault-diagnosis applications: model-based condition monitoring: actuators, drives, machinery, plants, sensors and fault-tolerant systems. Springer-Verlag Berlin Heidelberg, 1-354.).

Within the condition monitoring methods are those based on models (Camps Echevarría et al. (2014aCAMPS ECHEVARRÍA L, DE FRAGA CAMPOS VELHO H, BECCENERI J C, SILVA NETO A J & LLANES SANTIAGO O. 2014. The fault diagnosis inverse problem with Ant Colony Optimization and Ant Colony Optimization with dispersion. Appl Math Comput 227: 687-700.,bCAMPS ECHEVARRÍA L, LLANES SANTIAGO O, HERNÁNDEZ FAJARDO J A, SILVA NETO A J & JÍMENEZ SÁNCHEZ D. 2014a. A variant of the Particle Swarm Optimization for the improvement of fault diagnosis in industrial systems via faults estimation. Eng Appl Artif Intell 28: 36-51.), Ding (2008)DING S. 2008. Model-based Fault Diagnosis Techniques. Springer-Verlag Berlin Heidelberg, 1-473., Patan (2008PATAN K. 2008. Artificial neural networks for the modelling and fault diagnosis of technical processes. Springer-Verlag Berlin Heidelberg, 1-206.), Venkatasubramanian et al. (2003aVENKATASUBRAMANIAN V, RENGASWAMY R & KAVURI S N. 2003a. A Review of Process Fault Detection and Diagnosis, Part 1: Quantitative Model-Based Methods. Comput Chem Eng 27: 293-311.,bVENKATASUBRAMANIAN V, RENGASWAMY R & KAVURI S N. 2003b. A Review of Process Fault Detection and Diagnosis, Part 2: Qualitative models and search strategies. Comput Chem Eng 27: 313-326.)) and those based on historical data (Bernal de Lázaro et al. 2015BERNAL DE LÁZARO J M, PRIETO MORENO A, LLANES SANTIAGO O & SILVA NETO A J. 2015. Optimizing kernel methods to reduce dimensionality in fault diagnosis of industrial systems. Comput Ind Eng 87: 140-149., 2016BERNAL DE LÁZARO J M, LLANES-SANTIAGO O, PRIETO-MORENO A, KNUPP D & SILVA-NETO A J. 2016. Enhanced dynamic approach to improve the detection of small-magnitude faults. Chem Eng Sci 146: 166-179., Pang et al. 2014PANG Y Y, ZHU H P & LIU F M. 2014. Fault Diagnosis Method Based on KPCA and Selective Neural Network Ensemble. Adv Mater 915: 1272-1276., Sina et al. 2014SINA S, SADOUGH Z N & KHORASANI K. 2014. Dynamic neural network-based fault diagnosis of gas turbine engines. Neurocomputing 125: 153-165.). In the first approach, the use of models representing the operation of the processes is needed. The tools used in this approach are based on the generation of residuals obtained from the difference between the measurable signals from the real process and the values obtained from a model of this process. This implicates a high knowledge about the characteristics of the processes, their parameters, and operation zones. However, this knowledge is usually difficult to achieve due to the complexity of the industrial plants. On the other hand, the approaches based on historical data do not need a mathematical model, and they do not require much prior knowledge of the process parameters (Wang & Hu 2009WANG J & HU H. 2009. Vibration-based fault diagnosis of pump using fuzzy technique. Measurement 39: 176-185., Cerrada et al. 2016CERRADA M, SÁNCHEZ R, PACHECO F, CABRERA D, ZURITA G & LI C. 2016. Hierarchical feature selection based on relative dependency for gear fault diagnosis. Appl Intell 44: 687-703., 2018CERRADA M, SÁNCHEZ R, LI C, PACHECO F, CABRERA D, DE OLIVEIRA J V & VÁZQUEZ R E.2018. A review on data-driven fault severity assessment in rolling bearings. Mech Syst Signal Process 99: 169-196.). These characteristics constitute an advantage for complex systems, where relationships among variables are nonlinear, and not totally known. Therefore, it is difficult to obtain an analytical model that describes, efficiently, the dynamics of the process.

The use of techniques based on fuzzy tools has increased significantly in recent years in several scientific areas. In image processing the works Wang et al. (2020aWANG C, PEDRYCZ W, ZHOU M & LI Z. 2020. Sparse regularization-based Fuzzy C-Means clustering incorporating morphological grayscale reconstruction and wavelet frame. IEEE Trans. Fuzzy Syst 29(7): 1826-1840.,bWANG C, PEDRYCZ W, YANG J, ZHOU M & LI Z. 2020a. A Wavelet frame-based fuzzy C-means clustering for segmenting images on graphs. IEEE Trans Cybern 50(9): 3938-3949.) can be cited. In Gao et al. (2019)GAO S, ZHOU M, WANG Y, CHENG J, YACHI H. & WANG J. 2019. Dendritic neuron model with effective learning algorithms for classification, approximation and prediction. IEEE Trans Neural Netw Learn Syst 30: 601-614. a new dendritic neuron model (DNM) taking into account the nonlinearity of synapses, not only for a better understanding of a biological neuronal system, but also for providing a more useful method for solving, is proposed. In the field of control theory, the authors of Tong et al. (2019)TONG S, LIU W, QIAN D, YAN X & FANG J. 2019. Design of a networked tracking control system with a data-based approach. IEEE/CAA J Autom Sin 6: 1261-1267. propose a data-based design approach for a Networked Tracking Control System which utilizes the input-output data of the controlled process to establish a predictive model with the help of Fuzzy Cluster Modelling technology. In Liu et al. (2015)LIU L, YANG A, ZHOU W, ZHANG X, FEI M & TU X. 2015. Robust dataset classification approach based on neighbor searching and kernel fuzzy c-means. IEEE/CAA J Autom Sin 2: 235-247., a new robust dataset classification approach is presented based on neighbor searching and kernel fuzzy c-mean. The approach adopts the neighbor searching method with the dissimilarity matrix to normalize the dataset, and the number of clusters is determined by controlling clustering shape. In condition monitoring applications the works Rodríguez-Ramos et al. (2017RODRÍGUEZ-RAMOS A, LLANES-SANTIAGO O, BERNAL DE LÁZARO J, CRUZ-CORONA C, SILVA NETO A & VERDEGAY-GALDEANO J. 2017. A novel fault diagnosis scheme applying fuzzy clustering algorithms. Appl Soft Comput 58: 605-619., 2018bRODRÍGUEZ-RAMOS A, CRUZ-CORONA C, VERDEGAY J, SILVA NETO A & LLANES-SANTIAGO O. 2018. An approach for fault diagnosis using a novel hybrid fuzzy clustering algorithms. IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Rio de Janeiro, Brasil. 1-8.) can be cited.

A main aspect in the use of fuzzy sets is the provision of membership grades. In order to enhance the capability of fuzzy sets to capture and model user provided membership information, researchers have introduced non-standard second order fuzzy sets such as intuitionistic (Atanassov 1986ATANASSOV K. 1986. Intuitionistic fuzzy sets. Fuzzy Sets Syst. 20: 87-96., 2012ATANASSOV K. 2012. On Intuitionistic Fuzzy Sets Theory. Springer-Verlag Berlin Heidelberg, 1-324.) and interval type-2 fuzzy sets (Mendel et al. 2006MENDEL J M, JOHN R I, & LIU F. 2006. Interval type-2 fuzzy sets made simple. IEEE Trans Fuzzy Syst 14: 808-821., Mendel & Wu 2010MENDEL J M & WU D. 2010. Perceptual Computing. Wiley: Hoboken, 1-336.). These non–standard fuzzy sets allow the inclusion of imprecision and uncertainty in the specification of membership grades.

Recently, Prof. Ronald R. Yager introduced another class of non-standard fuzzy subset named Pythagorean fuzzy subset (Yager 2014)YAGER R R. 2014. Pythagorean Membership Grades in Multicriteria Decision Making. IEEE Trans Fuzzy Syst 22: 958-965.. In Yager (2014), it is shown that the space of Pythagorean membership grades is greater than the space of intuitionistic membership grades. This allows the use of the Pythagorean fuzzy sets in a greater set of applications than the intuitionistic fuzzy sets.

Normally, the data obtained from complex industrial processes are corrupted by noise. This introduces uncertainties in the observations which seriously affect the performance of the condition monitoring systems. On the one hand, this situation can cause false alarms when the fault diagnosis system confuses the Normal Operation Condition (NOC) with a fault. In another sense, the fault diagnosis system can present problems to correctly distinguish or classify a fault that is affecting the process. Both situations will lead to erroneous decision making, and may cause economic losses and affect industrial safety.

In order to overcome the problems mentioned above, and to obtain a robust condition monitoring scheme, an approach based on the use of Pythagorean membership grades is proposed which constitutes the main contribution of this paper. This approach allows to obtain a new variant of the known Fuzzy C-Means algorithm, called Pythagorean Fuzzy C-Mean algorithm (PyFCM), and it’s kernel version (KPyFCM) in order to achieve greater separability among classes and reduce classification errors. Both algorithms also constitute, other contributions of the paper.

The organization of the paper is as follow: in Section of Material and Methods, the general characteristics of the tools used in the proposed methodology are presented. Also, the principal theoretical aspects of the Pythagorean memberships grades theory are presented. In Section Description of the proposal, a description of the classification methodology using fuzzy clustering techniques is presented. In Section Study Cases and Experimental Design the proposed methodology is evaluated with four synthetic datasets and with the benchmark Tennessee Eastman (TE) process. Next, an analysis of the results obtained and a comparison with other computational tools is developed in Section Analysis of results. Finally, the conclusions are presented.

MATERIALS AND METHODS

In this section, a general description of Fuzzy C-Means (FCM), and Intuitionistic FCM (IFCM) algorithms are presented first. Later, the general characteristics of the Pythagorean membership grades, the new variant of the FCM algorithm, Pythagorean FCM algorithm (PyFCM), and its kernel version (KPyFCM) are also presented.

Fuzzy C-Means

Different methods have been proposed for fuzzy clustering. Among them, the most common are those based on distance. One of these methods, and the most popular, is the Fuzzy C-Means (FCM) algorithm (Bezdek et al. 1984BEZDEK J C, EHRLICH R & FULL W. 1984. FCM: The Fuzzy c-Means clustering algorithm. Comput Geosci 2-3: 191-203.) which uses the optimization criterion (1) to group the data according to the similarity between them.

J F C M = i = 1 c k = 1 N ( u i k ) m ( d i k ) 2 (1)

The exponent m>1 in (1), is an important factor that regulates the fuzziness of the resulting partition. The fuzzy clustering allows for obtaining the membership degrees matrix U=[uik]c×N where uik represents the fuzzy membership degree of the sample k to the ith class, which satisfies:

i = 1 c u i k = 1 , k = 1 , 2 , . . . , N (2)

where c is the number of classes and N is the number of samples. In this algorithm, the similarity is evaluated by using the distance function dik, represented by the Eq. (3). This function provides a measure of the distance between the data and the center of the classes v=v1,v2,...,vc, being 𝐀n×n the norm induction matrix, where n is the quantity of measured variables.

d i k 2 = ( 𝐱 k 𝐯 i ) T 𝐀 ( 𝐱 k 𝐯 i ) (3)

The measure of dissimilarity is the square distance between each data point and the clustering center 𝐯i. This distance is weighted by a power of the membership degree (uik)m. The value of the cost function J is a measure of the weighted total quadratic error and statistically, it can be seen as a measure of the total variance of 𝐱k regarding 𝐯i.

The conditions for local extreme in Eq. (1) and Eq. (2) are derived using Lagrangian multipliers (Bezdek 1981BEZDEK J. 1981. Pattern Recognition with Fuzzy Objective Function Algorithms. Plenum, New York.):

u i k = 1 j = 1 c ( d i k , 𝐀 / d j k , 𝐀 ) 2 / ( m 1 ) (4)
𝐯 i = k = 1 N ( u i k m 𝐱 k ) k = 1 N u i k m (5)

In Eq. (5), it should be noted that 𝐯i is the weighted average of the data elements that belong to a cluster, i.e., it is the center of the cluster i. FCM algorithm is an iterative procedure where N data are grouped in c classes. Initially, the user should establish the number of classes (c). The centers of the c classes are initialized in a random form, and they are modified during the iterative process. In a similar way, the membership degrees matrix U is modified until the convergence, i.e. UtUt1<ϵ, where ϵ is a tolerance limit prescribed a priori, and t is an iteration counter.

The stopping criteria used in this algorithm are:

  1. Criterion 1: Maximum number of iterations (Itrmax).

  2. Criterion 2: UtUt1<ϵ

Intuitionistic Fuzzy C-Means algorithm

Intuitionistic fuzzy c-means clustering algorithm (Chaira 2011CHAIRA T. 2011. A novel intuitionistic fuzzy c-means clustering algorithm and its application to medical images. Appl Soft Comput 11: 1711-1717.) is based upon intuitionistic fuzzy set theory. Fuzzy set generates only membership function μ(x),xX, whereas intuitionistic fuzzy set (IFS) given by Atanassov considers both membership μ(x) and nonmembership υ(x). An intuitionistic fuzzy set 𝐀 in X, is written as:

𝐀 = { x , μ 𝐀 ( x ) , υ 𝐀 ( x ) | x X } (6)

where μ𝐀(x)[0,1], υ𝐀(x)[0,1] are the membership and non-membership degrees of an element in the set 𝐀 with the condition: 0μ𝐀(x)+υ𝐀(x)1.

When υ𝐀(x)=1μ𝐀(x) for every x𝐀, then the set 𝐀 becomes a fuzzy set. For all intuitionistic fuzzy sets, a hesitation degree π𝐀(x) is also indicated (Atanassov 2012). It express the lack of knowledge in defining of whether x belongs to IFS or not and is given by:

π 𝐀 ( x ) = 1 μ 𝐀 ( x ) υ 𝐀 ( x ) ; 0 π 𝐀 ( x ) 1 (7)

A graphical representation of the meaning of π𝐀(x) is shown in Figure 1.


Graphical representation of the meaning of the hesitation degree 𝛑𝐀(𝐱).

Intuitionistic fuzzy c-means objective function contains two terms: (i) modified objective function of conventional FCM using Intuitionistic fuzzy set and (ii) intuitionistic fuzzy entropy (IFE). IFCM minimizes the objective function as:

J I F C M = i = 1 c k = 1 N u i k * m d i k 2 + i = 1 c π i * e 1 π * (8)

uik*=uikm+πik, where uik* denotes the intuitionistic fuzzy membership and uik denotes the conventional fuzzy membership of the kth data in the ith class. πik is the hesitation degree, which is defined as:

π i k = 1 u i k ( 1 u i k α ) 1 / α , α > 0 (9)

and it is calculated from Yager’s intuitionistic fuzzy complement as

N ( x ) = ( 1 x α ) 1 / α , α > 0 (10)

thus, with the help of Yager’s intuitionistic fuzzy complement, intuitionistic fuzzy set becomes:

A = { x , μ A ( x ) , ( 1 μ A ( x ) α ) 1 / α | x X } (11)

and

π i * = 1 N k = 1 N π i k , k [ 1 , N ] (12)

The second term in the objective function is called intuitionistic fuzzy entropy (IFE). Initially, the idea of fuzzy entropy was given by Zadeh in 1968 (Zadeh 1968ZADEH L. 1968. Probability measures of fuzzy events. J Math Anal Appl 23: 421-427.). It is the measure of fuzziness in a fuzzy set. Similarly in the case of IFS, intuitionistic fuzzy entropy gives the amount of vagueness or ambiguity in a set. For intuitionistic fuzzy cases, if μA(xi), vA(xi), πA(xi) are the membership, non-membership, and hesitation degrees of the elements of the set X=x1,x2,...,xn, then intuitionistic fuzzy entropy, IFE that denotes the degree of intuitionism in fuzzy set, may be given as:

I F E ( A ) = i = 1 n π A ( x i ) e [ 1 π A ( x i ) ] (13)

where πA(xi)=1μA(xi)vA(xi) IFE is introduced in the objective function to maximize the good points in the class. The goal is to minimize the entropy. Modified cluster centers are:

v i * = k = 1 n u i k * x k k = 1 n u i k * (14)

Pythagorean Fuzzy C-Means algorithm (PyFCM)

In , a new class of nonstandard fuzzy sets called Pythagorean fuzzy sets (PFS) is introduced. Hereinafter, the membership grades associated with these sets will be named as Pythagorean membership grades. Following, the main characteristics of the Pythagorean membership grades are presented.

For expressing the Pythagorean membership grades a pair of values r(x) and d(x) for each xX are assigned. Both values will be called as the strength of commitment at x in the case of r(x)[0,1] and the direction of commitment in the case of d(x)[0,1] . The values r(x) and d(x) are associated with a pair of membership grades AY(x) and AN(x). These memberships grades indicate the support for membership of x in A and the support against membership of x in A respectively. Next, it is shown that AY(x) and AN(x) are related using the Pythagorean complement with respect to r(x). More specially, the values of AY(x) and AN(x) are defined from r(x) and d(x) as

A Y ( x ) = r ( x ) c o s ( θ ( x ) ) (15)
A N ( x ) = r ( x ) s i n ( θ ( x ) ) (16)

where

θ ( x ) = ( 1 d ( x ) ) π 2 (17)

and θ(x)[0,π2] is expressed in radians.

First, it is shown that AY(x) and AN(x) are Pythagorean complements with respect to r(x). Squaring Eqs. (15) and (16)

A Y 2 ( x ) = r 2 ( x ) c o s 2 ( θ ( x ) ) (18)
A N 2 ( x ) = r 2 ( x ) s i n 2 ( θ ( x ) ) (19)

and, by adding both equations the following is obtained

A Y 2 ( x ) + A N 2 ( x ) = r 2 ( x ) ( c o s 2 ( θ ) + s i n 2 ( θ ) ) (20)

From the Pythagorean theorem, it is known that cos2(θ)+sin2(θ)=1. Then

A Y 2 ( x ) + A N 2 ( x ) = r 2 ( x ) ( c o s 2 ( θ ) + s i n 2 ( θ ) ) 1 (21)

and hence

A Y 2 ( x ) = r 2 ( x ) A N 2 ( x ) (22)

Thus, it is evident that AY and AN are Pythagorean complements with respect to r(x).

The direction of the strength, d(x), indicates on a scale of 1 to 0 how fully the strength r(x) is pointing to membership. From Eq. (17) θ(x) = (1d(x))π2. If d(x)=1, then θ(x)=0 , therefore, cos(θ(x))=1 and sin(θ(x))=0. From Eq. (16), AN(x)=0, and from Eq. (22) AY(x)=r(x). This indicates that the direction of r(x) is completely to membership. On the other hand, performing a similar analysis for d(x)=0, it is obtained that AY(x)=0 and AN(x)=r. This indicates that the direction of the strength is completely to nonmembership. Intermediate values of d(x) indicate partial support to membership and nonmembership.

In a general form, a Pythagorean membership grade is represented by a pair of values (a,b) such that a,b[0,1] and a2 + b2 1. In this case, a=AY(x), indicates the degree of support for membership of x in A and, b = AN(x) indicates the degree of support against membership of x in A. Taking into account the pair (a,b), the equation 17) can be expressed as a2 + b2 = r2. The latter indicates that a Pythagorean membership grade is a point of a circle of radius r.

An intuitionistic membership grade presented in is also a pair (a,b) that satisfies a,b[0,1] and a+b1. In ,it was demonstrated that the set of Pythagorean membership grades is greater than the set of intuitionistic membership grades. That result is clearly shown in Figure 2 taken from . Here, it is possible to observe that intuitionistic membership grades are all points under the line x + y 1 and the Pythagorean membership grades are all points with x2 + y2 1.

Figure 2
Comparison of space of Pythagorean and intuitionistic membership grades.

Taking into account the theory of Pythagorean fuzzy sets, it can be said that the objective function on the Pythagorean Fuzzy C-Means algorithm (PyFCM) is similar to the one obtained for the IFCM algorithm according equation (8). In this case, a hesitation degree, πA(x), is given by:

π A ( x ) = 1 μ A 2 ( x ) v A 2 ( x ) ; 0 π A ( x ) 1 (23)

Therefore, in equation (8), πik is defined as:

π i k = 1 u i k 2 ( 1 u i k α ) 2 / α , α > 0 (24)

The most important implication of this result is the possibility of using the Pythagorean fuzzy sets in a larger set of situations than intuitionistic fuzzy sets. In the case of fault diagnosis, this result allows to improve the classification process.

Pythagorean Fuzzy C-Means algorithm based on a kernel approach

KFCM represents the kernel version of FCM. This algorithm uses a kernel function for mapping the data points from the input space to a high dimensional space, as shown in Figure 3.

Figure 3
KFCM feature space and kernel space.

In this case, a kernel version of the PyFCM (KPyFCM) is obtained in order to achieve greater separability among classes, and reduce the classification errors. KPyFCM minimizes the objective function:

J K P y F C M = i = 1 c k = 1 N u i k * m 𝚽 ( 𝐱 𝐤 ) 𝚽 ( 𝐯 𝐢 ) 2 + i = 1 c π i * e 1 π * (25)

where uik*=uikm+πik, πik hesitation degree, which is defined according to Eq. (24) and πi* is defined as the equation (12).

Also, 𝚽(𝐱𝐤)𝚽(𝐯𝐢)2 is the square of the distance between 𝚽(𝐱𝐤) and 𝚽(𝐯𝐢). The distance in the feature space is calculated through the kernel in the input space as follows:

𝚽 ( 𝐱 𝐤 ) 𝚽 ( 𝐯 𝐢 ) 2 = 𝐊 ( 𝐱 𝐤 , 𝐱 𝐤 ) 2 𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) + 𝐊 ( 𝐯 𝐢 , 𝐯 𝐢 ) (26)

If the Gaussian kernel is used, then 𝐊(𝐱,𝐱)=1 and 𝚽(𝐱𝐤)𝚽(𝐯𝐢)2=2(1𝐊(𝐱𝐤,𝐯𝐢)). Thus, Eq. (25) can be written as:

J K P y F C M = 2 i = 1 c k = 1 N u i k * m 1 𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) 2 + i = 1 c π i * e 1 π * (27)

where,

𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) = e 𝐱 k 𝐯 i 2 / σ 2 (28)

It is possible to find many different kernel functions in the scientific literature, and the Gaussian kernel is one of the most popular. In general, the selection of a kernel depends on the application (Bernal de Lázaro et al. 2015, Motai 2015MOTAI Y. 2015. Kernel association for classification and prediction: A survey. IEEE Trans Neural Netw Learn Syst 26: 208-223., Nayak et al. 2015NAYAK J, NAIK B & BEHERA H. 2015. A comprehensive survey on support vector machine in data mining tasks: Applications & challenges. Int J Database Theory Appl 8: 169-186., 2016). In this paper, several experiments were performed using various kernel functions such as the Gaussian Kernel, the Polynomial Kernel and the Hyper-tangent Kernel. Considering the results that were obtained, the Gaussian Kernel was selected.

Minimizing Eq. (27) under the constraint shown in Eq. (2), yields:

u i k * = 1 j = 1 c ( 1 𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) 1 𝐊 ( 𝐱 𝐤 , 𝐯 𝐣 ) ) 1 / ( m 1 ) (29)
𝐯 i = k = 1 N ( u i k * m 𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) 𝐱 𝐤 ) k = 1 N u i k * m 𝐊 ( 𝐱 𝐤 , 𝐯 𝐢 ) (30)

KPyFCM algorithm is presented in Algorithm 1.

Description of the proposal

The classification scheme proposed in this paper is shown in Figure 4. It presents an offline training stage and an online recognition stage. In the training stage, the historical data of the process are used to train (modeling the functional stages through the clusters) a fuzzy classifier. After the training, the classifier is used online (recognition) in order to classify every new sample taken from the process. The result intends to offer information about the system state in real-time for the operator .

Figure 4
Classification scheme using fuzzy clustering.

The clustering methods group the data in different classes based on a measure of similitude. In the processes, the data are acquired by means of a SCADA (Supervisory Control and Data Acquisition) system, and the classes can be associated with functional states. In the case of statistical classifiers, each sample is compared with the center of each class by means of a measure of similitude to determine to which class the sample belongs. In the case of the fuzzy classifiers, the comparison is made to determine the membership degree of the sample to each class. In general, the higher membership degree determines the class to which the sample is assigned, as it is shown in (31).

C i = { i : m a x { μ i k } , i , k } (31)

Off-line training

In this stage, a historical dataset representative of the different operation states of the process (classes) is used for training the fuzzy classifier, where the center of each class is determined.

Online recognition

In this stage, the observations obtained by the SCADA system are classified one by one. In the classification process, the distance between the received observation and each one of the class centers is calculated. Next, the fuzzy membership degree of the observation to each one of the c classes is obtained. The observation will be assigned to the class with highest membership degree (See Algorithm 2).

The general condition monitoring scheme used for all performed experiments, is shown in Fig. 5. In this scheme, it can be seen that the on-line recognition algorithm determines the existence of a fault if j samples representative of it are received in a window of time. Later, an abnormal situation alarm of the process is executed. This is done with the aim of reducing false alarms in the presence of noise or outliers. The parameter j and the dimension of the window of time are selected by the expert operator of the plant in correspondence to the process.

Figure 5
Flowchart of the condition monitoring process.

Study cases and experimental design

In this section the datasets shown in Figure 6 are presented to validate the performance of the novel condition monitoring scheme proposed in this paper. These datasets were created synthetically to have complex situations for classification despite having only two classes of two variables. The datasets Data A, Data B and Data D have 1000 observations each one and the dataset Data C has 700 observations. These datasets will be identified in the rest of the paper as Experimental Datasets (ED). In the training, 750 observations from the Data A, B and D were used and 250 observations were used in the recognition stage. In the case of the Data C, 525 observations were used in the training and 175 observation were used in the recognition stage.

Several experiments (σ=10,20,30,40,…,100) were performed and the value of σ that gave the best results in the classification was selected.

The values of the parameters used for the applied algorithms were: Number of iterations = 100, ϵ = 105, m = 2, σ = 10 (only used for the version kernel of the algorithms).

Figure 6
Datasets for experiments.

The second case study is the Tennessee Eastman (TE) process benchmark which has been widely used to evaluate the performance of new control and monitoring strategies (Yin et al. 2012YIN S, DING S X, HAGHANI A, HAO H & ZHANG P. 2012. A comparison study of basic data-driven fault diagnosis and process monitoring methods on the benchmark Tennessee Eastman process. J Process Control 22(9): 1567-1581., Prieto-Moreno et al. 2015PRIETO-MORENO A, LLANES-SANTIAGO O & GARCÍA-MORENO E. 2015. Principal components selection for dimensionality reduction using discriminat information applied to fault diagnosis. J Process Control 33: 14-24., Llanes-Santiago et al. 2019LLANES-SANTIAGO O, RIVERO-BENEDICO B, GÁLVEZ-VIERA S, RODRÍGUEZ-MORANT R, TORRES-CABEZA R & SILVA NETO A. 2019. A fault diagnosis proposal with online imputation to incomplete observations in industrial plants. Rev Mex Ing Quim 18(1): 83-98.). The process consists of five major units interconnected as shown in Figure 7.

Figure 7
Piping diagram of the Tennessee Eastman process.

This benchmark contains 21 preprogrammed faults and one normal operating condition dataset. The datasets of the TE are generated for 48h with the inclusion of faults after 8 hours of simulation. The control objectives and general features of the process simulation are described in the paper . All data sets used in this paper can be downloaded from http://web.mit.edu/braatzgroup/TE_process.zip. Table 1 shows the faults considered to evaluate the benefits of the proposal presented in this paper. For the training, 480 observations of each fault were used, and 960 observations in the online recognition.

Another problem of the current fuzzy clustering method is related to the correct selection of its parameters which is decisive in obtaining a high performance. Nowadays, these issues of crucial importance are open problems in the fault diagnosis applications and in others research fields (Wang et al. 2020a, Filho et al. 2015FILHO T S, PIMENTEL B, SOUZA R & OLIVEIRA A. 2015. Hybrid methods for fuzzy clustering based on fuzzy c-means and improved particle swarm optimization. Expert Syst Appl 42: 6315-6328., 2016GAO S, ZHOU M, WANG Y, CHENG J, YACHI H. & WANG J. 2019. Dendritic neuron model with effective learning algorithms for classification, approximation and prediction. IEEE Trans Neural Netw Learn Syst 30: 601-614.). The parameters Number of iterations, ϵ, m and σ were selected according to the experience in previous works (Rodríguez-Ramos et al. 2019RODRÍGUEZ-RAMOS A, SILVA NETO A J & LLANES-SANTIAGO O. 2019. Fault Detection Using Kernel Computational Intelligence Algorithm Computational Intelligence. Optimization and Inverse Problems with Applications in Engineering. Springer. 63-75., 2018aRODRÍGUEZ-RAMOS A, SILVA NETO A J & LLANES-SANTIAGO O. 2018. An approach to fault diagnosis with online detection of novel faults using fuzzy clustering tools. Expert Syst Appl 113: 200-212.).

The values of the parameters used for the applied algorithms are: Number of iterations = 100, ϵ = 105, m = 2, σ = 50 (only used for the version kernel of the algorithm).

Table I
Description of faults of the TE process.

Analysis of results

A very important step in the design of fault diagnosis systems consist in verifying the quality of the performed task. The most used criterion for this analysis is the confusion matrix (CM). The confusion matrix is an indicator that allows for obtaining the performance of the classifier in the classification process. Each CMrs element of a confusion matrix for rs, indicates the number of times that the classifier confuses a state r with a state s in a set of L experiments. The results obtained from the application of the proposed methodology to fault diagnosis by using the experimental datasets (ED) and in the TE process are presented next. Figure 8 illustrates a confusion matrix and the respective interpretation.

Figure 8
Confusion matrix interpretation for a binary classification process (Bernal de Lázaro et al. 2016).

Note that the main diagonal of the matrix CMrs represents the number of correct samples detected/classified and the values out of this diagonal reflect the confusion between the classes or the operation states.

Results for the experimental datasets

Tables 2, 3, 4 and 5 show the confusion matrix for the experimental dataset (ED) where C1: Class 1 and C2: Class 2 (C1, C2: 250 (for Data A, Data B and Data D), C1, C2: 175 (for Data C)). The main diagonal is associated with the number of observations successfully classified. Since the total number of observations per class is known, the accuracy (TA=correctly classified observations/total observations) can also be computed. The last row shows the average (AVE) of TA.

Table II
CM for the experimental data set (Data A).
Table III
CM for the experimental data set (Data B).
Table IV
CM for the experimental data set (Data C).
Table V
CM for the experimental data set (Data D).

Figure 9 shows the classification results by using the FCM, IFCM, PyFCM, KFCM, KIFCM and KPyFCM algorithms. That shows a global classification percentage obtained for each algorithm.

Figure 9
Classification for experimental datasets

Results for TE process

Table 6 shows the confusion matrix for experimental dataset where F1: Fault 1, F2: Fault 2, F6: Fault 6 and F7: Fault 7.

Table VI
CM for the TE process (F1: 960, F2: 960, F6: 960, F7: 960).

Figure 10 shows the classification results for the faults 1, 2, 6 and 7 by using the FCM, IFCM, PyFCM, KFCM, KIFCM and KPyFCM algorithms for TE process. A summary of the results can be seen in Figure >11, that shows a global classification percentage obtained for each algorithm.

Figure 10
Faults classification (%) for the TE process.
Figure 11
Global classification (%) obtained for each algorithm for the TE process.

All experiments were performed on a computer with the following characteristics: Intel Core i7-6500U 2.5 - 3.1GHz, memory: 8GB DDR3L. The Table 7 shows that the average time delayed each algorithm to perform an execution. When comparing these execution times with the time constant of the TE process, it can be seen, that they are very small and therefore show the feasibility of applying these algorithms in the classification scheme.

Table VII
Analysis of computational time.

As several algorithms are presented, it is necessary to analyze if there are significant differences among the results of them. To achieve this, it is necessary to apply statistical tests (García amp; Herrera 2008GARCÍA S, MOLINA D, LOZANO M & HERRERA F. 2009. A study on the use of non-parametric tests for analyzing the evolutionary algorithms behaviour: a case study on the CEC’2005 Special Session on Real Parameter Optimization. J Heuristics 15: 617-644., García et al. 2009GARCÍA S & HERRERA F. 2008. An Extension on Statistical Comparisons of Classifiers over Multiple Data Sets for all Pairwise Comparisons. J Mach Learn Res 9: 2677-2694., Luengo et al. 2009LUENGO J, GARCÍA S & HERRERA F. 2009. A study on the use of statistical tests for experimentation with neural networks: Analysis of parametric test conditions and non-parametric tests. Expert Syst Appl 36: 7798-7808.).

Statistical tests

First, the non-parametric Friedman test is applied in order to demonstrate that there is at least one algorithm whose results have significant differences with respect to the results of the others. Afterward, if the null-hypothesis of the Friedman test is rejected, it is necessary to make a comparison in pairs to determine the best algorithm(s). For this, the non-parametric Wilcoxon test is applied.

Friedman Test

In our case, for six experiments (k=6) and 10 datasets (N=10), the value of statistical Friedman FF = 340 was obtained. With k=6 and N=10, FF is distributed according to the F distribution with 61=5 and (61)×(101)=45 degrees of freedom. The critical value of F(5,45) for α=0.05 is 2.4221, so the null-hypothesis is rejected (F(5,45) < FF) which means that at least the average performance of at least one algorithm is significantly different from the average value of the performance of other algorithms.

Wilcoxon Test

Table 8 shows the results of the comparison in pairs of the algorithms (1: FCM, 2: IFCM, 3: PyFCM, 4: KFCM, 5: KIFCM, 6: KPyFCM) using the Wilcoxon test. The first two rows contain the values of the sum of the positive (R+) and negative (R) rank for each comparison established. The next two rows show the statistical values T and the critical value of T for a level of significance α=0.05. The last row indicates which algorithm was the winner in each comparison. The summary in Table 8 shows the number of times that each algorithm was the winner.

Table VIII
Result of the comparison between the experiments.

As can be seen, among the FCM, IFCM and PyFCM algorithms, the PyFCM algorithm obtains the better results. In the analysis with the Kernel algorithms, the KPyFCM algorithm obtains the better results. Taking into account all algorithms, it is shown that the KPyFCM algorithm obtains the best results.

CONCLUSIONS

The main contribution of this paper is the development of a robust scheme for condition monitoring in industrial systems by using Pythagorean membership grades. The fundamental motivation for this proposal is based on the fact that the space of Pythagorean membership grades is greater than the space of the standard and intuitionistic membership grades. In the classification process, an observation is assigned to the class in which it achieves the highest membership degree. Pythagorean membership functions allow for the use of a larger set of numeric values for assigning the membership degree to an observation; and that a subset of them have a numeric value greater than those allowed by institutionistic membership functions. In the classification process, the membership degree to a class is maximized. If there is a greater number of values in the search space and these values are greater than the institutionistic case, then this allows for the improvement of the classification process as can be seen in two case studies in the paper .

In the proposal, the FCM algorithm is modified by using Pythagorean fuzzy sets, and a new variant of that algorithm called Pythagorean Fuzzy C-Mean (PyFCM) algorithm is obtained. In addition, a kernel version of the PFCM algorithm (KPyFCM) is obtained in order to achieve greater separability among the classes, for reducing the classification errors. The approach proposed was validated using synthetic datasets and the TE process benchmark. The promising results obtained indicate the feasibility of the proposal.

ACKNOWLEDGMENTS

Authors acknowledge the financial support provided by the project TIN2017-86647-P from the Spanish Ministry of Economy and Competitivenes, including FEDER funds from the European Union; Fundacão Carlos Chagas Filho de Amparo à Pesquisa do Estado do Rio de Janeiro (FAPERJ); Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq); Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES) (Finance Code 001) provided by the project CAPES-PRINT Process No. 88881.311758/2018-01, research supporting agencies from Brazil; and CUJAE, Universidad Tecnológica de La Habana José Antonio Echeverría from Cuba.

REFERENCES

  • ATANASSOV K. 1986. Intuitionistic fuzzy sets. Fuzzy Sets Syst. 20: 87-96.
  • ATANASSOV K. 2012. On Intuitionistic Fuzzy Sets Theory. Springer-Verlag Berlin Heidelberg, 1-324.
  • BERNAL DE LÁZARO J M, PRIETO MORENO A, LLANES SANTIAGO O & SILVA NETO A J. 2015. Optimizing kernel methods to reduce dimensionality in fault diagnosis of industrial systems. Comput Ind Eng 87: 140-149.
  • BERNAL DE LÁZARO J M, LLANES-SANTIAGO O, PRIETO-MORENO A, KNUPP D & SILVA-NETO A J. 2016. Enhanced dynamic approach to improve the detection of small-magnitude faults. Chem Eng Sci 146: 166-179.
  • BEZDEK J. 1981. Pattern Recognition with Fuzzy Objective Function Algorithms. Plenum, New York.
  • BEZDEK J C, EHRLICH R & FULL W. 1984. FCM: The Fuzzy c-Means clustering algorithm. Comput Geosci 2-3: 191-203.
  • CAMPS ECHEVARRÍA L, DE FRAGA CAMPOS VELHO H, BECCENERI J C, SILVA NETO A J & LLANES SANTIAGO O. 2014. The fault diagnosis inverse problem with Ant Colony Optimization and Ant Colony Optimization with dispersion. Appl Math Comput 227: 687-700.
  • CAMPS ECHEVARRÍA L, LLANES SANTIAGO O, HERNÁNDEZ FAJARDO J A, SILVA NETO A J & JÍMENEZ SÁNCHEZ D. 2014a. A variant of the Particle Swarm Optimization for the improvement of fault diagnosis in industrial systems via faults estimation. Eng Appl Artif Intell 28: 36-51.
  • CERRADA M, SÁNCHEZ R, PACHECO F, CABRERA D, ZURITA G & LI C. 2016. Hierarchical feature selection based on relative dependency for gear fault diagnosis. Appl Intell 44: 687-703.
  • CERRADA M, SÁNCHEZ R, LI C, PACHECO F, CABRERA D, DE OLIVEIRA J V & VÁZQUEZ R E.2018. A review on data-driven fault severity assessment in rolling bearings. Mech Syst Signal Process 99: 169-196.
  • CHAIRA T. 2011. A novel intuitionistic fuzzy c-means clustering algorithm and its application to medical images. Appl Soft Comput 11: 1711-1717.
  • DING S. 2008. Model-based Fault Diagnosis Techniques. Springer-Verlag Berlin Heidelberg, 1-473.
  • DOWNS J & VOGEL E. 1993. A plant wide industrial process problem control. Comput Chem Eng 17(3): 245-255.
  • FILHO T S, PIMENTEL B, SOUZA R & OLIVEIRA A. 2015. Hybrid methods for fuzzy clustering based on fuzzy c-means and improved particle swarm optimization. Expert Syst Appl 42: 6315-6328.
  • GAO S, ZHOU M, WANG Y, CHENG J, YACHI H. & WANG J. 2019. Dendritic neuron model with effective learning algorithms for classification, approximation and prediction. IEEE Trans Neural Netw Learn Syst 30: 601-614.
  • GARCÍA S, MOLINA D, LOZANO M & HERRERA F. 2009. A study on the use of non-parametric tests for analyzing the evolutionary algorithms behaviour: a case study on the CEC’2005 Special Session on Real Parameter Optimization. J Heuristics 15: 617-644.
  • GARCÍA S & HERRERA F. 2008. An Extension on Statistical Comparisons of Classifiers over Multiple Data Sets for all Pairwise Comparisons. J Mach Learn Res 9: 2677-2694.
  • HWANG I, KIM S, KIM Y & SEAH C. 2010. A Survey of Fault Detection, Isolation, and Reconfiguration Methods. IEEE Control Syst 18: 636-656.
  • ISERMANN R. 2011. Fault-diagnosis applications: model-based condition monitoring: actuators, drives, machinery, plants, sensors and fault-tolerant systems. Springer-Verlag Berlin Heidelberg, 1-354.
  • LIU L, YANG A, ZHOU W, ZHANG X, FEI M & TU X. 2015. Robust dataset classification approach based on neighbor searching and kernel fuzzy c-means. IEEE/CAA J Autom Sin 2: 235-247.
  • LLANES-SANTIAGO O, RIVERO-BENEDICO B, GÁLVEZ-VIERA S, RODRÍGUEZ-MORANT R, TORRES-CABEZA R & SILVA NETO A. 2019. A fault diagnosis proposal with online imputation to incomplete observations in industrial plants. Rev Mex Ing Quim 18(1): 83-98.
  • LUENGO J, GARCÍA S & HERRERA F. 2009. A study on the use of statistical tests for experimentation with neural networks: Analysis of parametric test conditions and non-parametric tests. Expert Syst Appl 36: 7798-7808.
  • MENDEL J M, JOHN R I, & LIU F. 2006. Interval type-2 fuzzy sets made simple. IEEE Trans Fuzzy Syst 14: 808-821.
  • MENDEL J M & WU D. 2010. Perceptual Computing. Wiley: Hoboken, 1-336.
  • MOTAI Y. 2015. Kernel association for classification and prediction: A survey. IEEE Trans Neural Netw Learn Syst 26: 208-223.
  • NAYAK J, NAIK B & BEHERA H. 2015. A comprehensive survey on support vector machine in data mining tasks: Applications & challenges. Int J Database Theory Appl 8: 169-186.
  • PANG Y Y, ZHU H P & LIU F M. 2014. Fault Diagnosis Method Based on KPCA and Selective Neural Network Ensemble. Adv Mater 915: 1272-1276.
  • PATAN K. 2008. Artificial neural networks for the modelling and fault diagnosis of technical processes. Springer-Verlag Berlin Heidelberg, 1-206.
  • PRIETO-MORENO A, LLANES-SANTIAGO O & GARCÍA-MORENO E. 2015. Principal components selection for dimensionality reduction using discriminat information applied to fault diagnosis. J Process Control 33: 14-24.
  • RODRÍGUEZ-RAMOS A, LLANES-SANTIAGO O, BERNAL DE LÁZARO J, CRUZ-CORONA C, SILVA NETO A & VERDEGAY-GALDEANO J. 2017. A novel fault diagnosis scheme applying fuzzy clustering algorithms. Appl Soft Comput 58: 605-619.
  • RODRÍGUEZ-RAMOS A, SILVA NETO A J & LLANES-SANTIAGO O. 2018. An approach to fault diagnosis with online detection of novel faults using fuzzy clustering tools. Expert Syst Appl 113: 200-212.
  • RODRÍGUEZ-RAMOS A, CRUZ-CORONA C, VERDEGAY J, SILVA NETO A & LLANES-SANTIAGO O. 2018. An approach for fault diagnosis using a novel hybrid fuzzy clustering algorithms. IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Rio de Janeiro, Brasil. 1-8.
  • RODRÍGUEZ-RAMOS A, SILVA NETO A J & LLANES-SANTIAGO O. 2019. Fault Detection Using Kernel Computational Intelligence Algorithm Computational Intelligence. Optimization and Inverse Problems with Applications in Engineering. Springer. 63-75.
  • SINA S, SADOUGH Z N & KHORASANI K. 2014. Dynamic neural network-based fault diagnosis of gas turbine engines. Neurocomputing 125: 153-165.
  • TONG S, LIU W, QIAN D, YAN X & FANG J. 2019. Design of a networked tracking control system with a data-based approach. IEEE/CAA J Autom Sin 6: 1261-1267.
  • VENKATASUBRAMANIAN V, RENGASWAMY R & KAVURI S N. 2003a. A Review of Process Fault Detection and Diagnosis, Part 1: Quantitative Model-Based Methods. Comput Chem Eng 27: 293-311.
  • VENKATASUBRAMANIAN V, RENGASWAMY R & KAVURI S N. 2003b. A Review of Process Fault Detection and Diagnosis, Part 2: Qualitative models and search strategies. Comput Chem Eng 27: 313-326.
  • WANG J & HU H. 2009. Vibration-based fault diagnosis of pump using fuzzy technique. Measurement 39: 176-185.
  • WANG C, PEDRYCZ W, ZHOU M & LI Z. 2020. Sparse regularization-based Fuzzy C-Means clustering incorporating morphological grayscale reconstruction and wavelet frame. IEEE Trans. Fuzzy Syst 29(7): 1826-1840.
  • WANG C, PEDRYCZ W, YANG J, ZHOU M & LI Z. 2020a. A Wavelet frame-based fuzzy C-means clustering for segmenting images on graphs. IEEE Trans Cybern 50(9): 3938-3949.
  • YAGER R R. 2014. Pythagorean Membership Grades in Multicriteria Decision Making. IEEE Trans Fuzzy Syst 22: 958-965.
  • YIN S, DING S X, HAGHANI A, HAO H & ZHANG P. 2012. A comparison study of basic data-driven fault diagnosis and process monitoring methods on the benchmark Tennessee Eastman process. J Process Control 22(9): 1567-1581.
  • ZADEH L. 1968. Probability measures of fuzzy events. J Math Anal Appl 23: 421-427.

Publication Dates

  • Publication in this collection
    05 Dec 2022
  • Date of issue
    2022

History

  • Received
    30 Apr 2020
  • Accepted
    22 Mar 2021
Academia Brasileira de Ciências Rua Anfilófio de Carvalho, 29, 3º andar, 20030-060 Rio de Janeiro RJ Brasil, Tel: +55 21 3907-8100 - Rio de Janeiro - RJ - Brazil
E-mail: aabc@abc.org.br