Skip to content
BY-NC-ND 3.0 license Open Access Published by De Gruyter July 6, 2017

An Improved Correlation Coefficient of Intuitionistic Fuzzy Sets

  • Han-Liang Huang EMAIL logo and Yuting Guo

Abstract

The intuitionistic fuzzy set is a useful tool to deal with vagueness and uncertainty. Correlation coefficient of the intuitionistic fuzzy sets is an important measure in intuitionistic fuzzy set theory and has great practical potential in a variety of areas, such as decision making, medical diagnosis, pattern recognition, etc. In this paper, an improved correlation coefficient of the intuitionistic fuzzy sets is defined, and it can overcome some drawbacks of the existing ones. The properties of this correlation coefficient are discussed. Then, the generalization of the coefficient of interval-valued intuitionistic fuzzy sets is also introduced. Finally, two examples about the application of the proposed correlation coefficient of the intuitionistic fuzzy sets in medical diagnosis and clustering are shown to illustrate the advantages over the existing methods.

2010 AMS Classification: 03E72

1 Introduction

The concept of the intuitionistic fuzzy set (IFS) proposed by Atanassov [1] is a generalization of the fuzzy set [26]. One of the characterization of the IFS is that it assigns to each element a membership degree and a non-membership degree rather than the membership degree only. The IFS is more flexible and practical for dealing with vagueness and uncertainty than the ordinary fuzzy set in many real situations. Then, Atanassov and Gargov [2] further generalized the IFS to an interval-valued intuitionistic fuzzy set (IVIFS) in which the values of its membership degree and non-membership degree are intervals. The IFS is closely related to other generalized fuzzy sets such as L-fuzzy sets [5, 7, 18, 23] and interval-valued fuzzy sets (IVFSs) [13, 15, 27, 29], and it has been widely applied in medical diagnosis [8, 14, 30], decision making [20, 24], clustering [3, 9, 25] and pattern recognition [4, 6, 12].

The measurement of correlation between two IFSs plays an important role in IFS theory, and it is always formulated by correlation coefficient. Therefore, how to define an effective correlation coefficient formula is an interesting research topic. Many scholars have paid great attention to this issue and obtained many useful results. For example, Hung and Wu [11] proposed a method to calculate the correlation coefficient of IFSs by means of “centroid”, which reflected not only the correlation between IFSs but also their positive or negative correlation. Moreover, they extended the “centroid” method to IVIFSs. Hung [10] investigated the correlation measure of IFSs from the perspective of statistics. Xu developed a correlation coefficient formula [20] of IFSs, which was generalized to the IVIFSs. Then, Xu proposed a decision-making method in medical diagnosis under the IFS environment. He also proposed another correlation coefficient formula in Ref. [21], which took the membership, non-membership, and hesitancy into account simultaneously.

The correlation coefficient formulas proposed by Xu [19, 21] may have some limitations. First of all, the formulas [see Eqs. (1) and (2)] are expressed by the form of quotient, which is of 0/0 type if two IFSs are the same. Moreover, the results obtained by Xu’s correlation coefficients are not coincident with our intuition in some situations. For example, the formulas may be equal to 1 even if A1A2 (see example 1).

In this paper, we improve Xu’s correlation coefficients so that it can overcome the drawbacks above. Two examples can show the effectivity of our proposed method. In Section 3, we introduce a new formula for calculating the correlation between IFSs and summarize the advantage of the improved correlation coefficient by comparing with Xu’s correlation coefficients [19, 21]. We also generalize it to IVIFS environment. Furthermore, we show some applications of the improved correlation coefficient in medical diagnosis and clustering in Section 4.

2 Preliminaries

In this section, some basic concepts of the IFSs and IVIFSs are presented.

Definition 1 ([1]): Let X be a universe of discourse, then an IFS A is defined as:

A={(x,μA(x),νA(x))|xX},

where μA(x):X→[0, 1] and νA(x):X→[0, 1], with 0≤μA(x)+νA(x)≤1. The numbers μA(x) and νA(x) represent the membership degree and the non-membership degree of the element x to the set A, respectively. For each IFS A over X, if πA(x)=1−μA(x)−νA(x), xX, then πA(x) is called the hesitation degree of the element x to the set A.

Let IFS(X) be the set of all IFSs over X. Then for A1, A2IFS(X), we have:

  1. A1A2 if and only if μA1(x)μA2(x) and νA1(x)νA2(x) for each xX;

  2. A1=A2 if and only if A1A2 and A2A1.

Definition 2 ([2]): Let X be a universe of discourse, then an IVIFS Ã is defined as:

A˜={(x,μ˜A˜(x),ν˜A˜(x))|xX},

where μ˜A˜(x)=[μ˜A˜L(x),μ˜A˜U(x)][0,1] and ν˜A˜(x)=[ν˜A˜L(x),ν˜A˜U(x)][0,1] are intervals, and 0μ˜A˜U(x)+ν˜A˜U(x)1 for each xX.

Especially, if μ˜A˜L(x)=μ˜A˜U(x) and ν˜A˜L(x)=ν˜A˜U(x), then the IVIFS Ã is reduced to an ordinary IFS.

Let IVIFS(X) be the set of all IVIFSs over X. Then for Ã1, Ã2IVIFS(X), we have:

  1. Ã1Ã2 if and only if μ˜A˜1U(x)μ˜A˜2U(x),μ˜A˜1L(x)μ˜A˜2L(x),ν˜A˜1U(x)ν˜A˜2U(x) and ν˜A˜1L(x)ν˜A˜2L(x) for each xX;

  2. Ã1=Ã2 if and only if Ã1Ã2 and Ã2Ã1.

  3. The correlation coefficient of two IFSs is introduced by Xu as follows.

Definition 3 ([19]): Let A1, A2IFS(X), and X={x1, x2, …, xn} be a finite universe of discourse, then we define

(1)ρ1(A1,A2)=12ni=1n(Δμmin+ΔμmaxΔμi+Δμmax+Δνmin+ΔνmaxΔνi+Δνmax)

as a correlation coefficient of the IFSs A1 and A2, where

Δμi=|μA1(xi)μA2(xi)|,   Δνi=|νA1(xi)νA2(xi)|(i=1,2,,n),Δμmin=mini{|μA1(xi)μA2(xi)|},   Δνmin=mini{|νA1(xi)νA2(xi)|},Δμmax=maxi{|μA1(xi)μA2(xi)|},   Δνmax=maxi{|νA1(xi)νA2(xi)|}.

IFS contains three elements, such as membership, non-membership, and hesitation. However, the above correlation coefficient of Definition 3 does not take the hesitation into account. Thus, Xu improved it in Ref. [21]:

(2)ρ2(A1,A2)=13ni=1n(Δμmin+ΔμmaxΔμi+Δμmax+Δνmin+ΔνmaxΔνi+Δνmax+Δπmin+ΔπmaxΔπi+Δπmax),

where

πA1(xi)=1μA1(xi)νA1(xi),   πA2(xi)=1μA2(xi)νA2(xi),Δπi=|πA1(xi)πA2(xi)|,   Δπmin=mini{|πA1(xi)πA2(xi)|},Δπmax=maxi{|πA1(xi)πA2(xi)|}(i=1,2,,n).

Remark 1: The above correlation coefficients ρk(Ai, Aj) (k=1, 2) satisfy the following properties:

  1. 0≤ρk(Ai, Aj)≤1;

  2. ρk(Ai, Aj)=ρk(Aj, Ai) (i, j=1, 2, …, m).

Clustering analysis is an important application of correlation coefficient. In what follows, we recall some concepts related to clustering analysis such as association matrix and equivalent association matrix.

Definition 4 ([22]): Let AjIFS(X), (j=1, 2, …, m), then C=(cij)m×m is called an association matrix, where cij=ρ(Ai, Aj) (i, j=1, 2, …, m) is the correlation coefficient of Ai and Aj, which have the following properties:

  1. 0≤cij≤1<mspace;(i, j=1, 2, …, m);

  2. cij=cji (i, j=1, 2, …m);

  3. cij=1<mspace;if and only if Ai=Aj.

Definition 5 ([22]): Let C=(cij)m×m be an association matrix, if

C2=CC=(c¯ij)m×m,

then C2 is called a composition matrix of C, where c¯ij=maxk{min{cik,ckj}} (i, j=1, 2, …, m).

Definition 6 ([22]): Let C=(cij)m×m be an association matrix, and C2 be its composition matrix, if C2C, i.e.,

maxk{min{cik,ckj}}cij

for all i, j=1, 2, …, m, then C is called an equivalent association matrix.

By the transitivity principle of equivalent matrix [17], it is easy to prove the following:

Theorem 1 ([22]): Let C=(cij)m×m be an association matrix, then after the finite times of compositions:

CC2C4C2k,

there must exist a positive integer k such that C2k=C2(k+1), and C2k is an equivalent association matrix.

Definition 7 ([22]): Let C=(cij)m×m be an equivalent association matrix, then we call Cλ =(λcij)m×m the λ-cutting matrix of C, where

λcij={0,cij<λ;1,cijλ.(i,j=1,2,,m)

and λ is the confidence level with λ∈[0, 1].

3 Main Results

In this section, we will introduce an improved correlation coefficient and generalize it to IVIFS environment.

The correlation coefficients proposed by Xu [20, 21] share the two properties of Remark 1. However, both the correlation coefficient formulas have the following limitations:

  1. The quotients in formulas (1) and (2) are of the 0/0 type if the formulas satisfy A1=A2, and it is unmeaningful in mathematical logic. Although we can deliberately set a special value for this 0/0 type quotient, it is more reasonable to avoid emerging this case.

  2. Formula (1) as well as the improved formula (2) may be equal to 1 even if A1A2. Hence, neither Formula (1) nor (2) satisfies the condition: ρ(Ai, Aj)=1⇔Ai=Aj, then the two formulas cannot be applied in clustering sometimes. Let us employ an example to illustrate our idea.

Example 1: Let A1 and A2 be two IFSs in X={x1, x2} given by

A1={x1,0.4,0.3,x2,0.3,0.2};A2={x1,0.3,0.2,x2,0.2,0.1}.

Obviously, A1A2. However, ρ1(A1, A2)=1 and ρ2(A1, A2)=1.

Thus, how to derive the correlation coefficients of IFSs satisfying the desirable property above (ρ(Ai, Aj)=1⇔Ai=Aj) is an interesting research topic. In order to solve this problem, we develop a new definition of correlation coefficient of the IFSs.

3.1 An Improved Correlation Coefficient of the IFSs

In the following, we define an improved correlation coefficient of the IFSs.

Definition 8: Let A1, A2IFS(X), and X={x1, x2, …, xn} be a finite universe of discourse, then we define

(3)ρ(A1,A2)=12ni=1n[αi(1Δμi)+βi(1Δνi)]

as a correlation coefficient of the IFSs A1 and A2, where

αi=cΔμiΔμmaxcΔμminΔμmax,βi=cΔνiΔνmaxcΔνminΔνmax(c>2,i=1,2,,n),Δμi=|μA1(xi)μA2(xi)|,   Δνi=|νA1(xi)νA2(xi)|,Δμmin=mini{|μA1(xi)μA2(xi)|},   Δνmin=mini{|νA1(xi)νA2(xi)|},Δμmax=maxi{|μA1(xi)μA2(xi)|},   Δνmax=maxi{|νA1(xi)νA2(xi)|}.

Remark 2: Note that the condition “c>2” ensures that both αi and βi belong to (0, 1). Hence, the correlation coefficient ρ*(A1, A2) satisfies the condition 0≤ρ*(A1, A2)≤1. If some αi and βi are <0, it will cause ρ*(A1, A2)<0. We can see example 2 as follows.

Example 2: Let A1 and A2 be two IFSs in X={x1, x2} given by

A1={x1,0.9,0.1,x2,0.7,0.2};A2={x1,0.1,0.8,x2,0.6,0.4}.

By Eq. (3), we can have α1=−6, α2=1, β1=−4, β2=1 when c=1, then we can calculate that ρ1(A1, A2) =−0.175<0.

We present some properties of the improved correlation coefficient in the following.

Theorem 2:The correlation coefficient ρ*(A1, A2) satisfies the following properties:

  1. ρ*(A1, A2)=ρ*(A2, A1);

  2. 0≤ρ*(A1, A2)≤1;

  3. A1=A2ρ*(A1, A2)=1.

Proof

  1. Obviously.

  2. Since 0<αi≤1, 0<βi≤1 and 0≤1−Δμi≤1, 0≤1−Δνi≤1, then

    0αi(1Δμi)+βi(1Δνi)2(i=1,2,,n),

    and thus, by Eq. (3), we know that 0≤ρ*(A1, A2)≤1.

  3. “⇒” If A1=A2, it implies that: μA1(xi)=μA2(xi),νA1(xi)=νA2(xi) for all xiX (i=1, 2, …, n), then

    Δμi=Δμmin=Δμmax=0,Δνi=Δνmin=Δνmax=0.

Thus, ρ*(A1, A2)=1.

“⇐” If ρ*(A1, A2)=1, by the fact that

0αi(1Δμi)1,0βi(1Δνi)1,

we have

αi(1Δμi)+βi(1Δνi)=2 and αi(1Δμi)=βi(1Δνi)=1(i=1,2,,n).

Since

0<αi1,   0<βi1,   01Δμi1,   01Δνi1,

then we obtain

αi=βi=1,1Δμi=1,1Δνi=1,i.e.,Δμi=Δνi=0.

Thus, μA1(xi)=μA2(xi),νA1(xi)=νA2(xi) for all xiX, i.e., A1=A2.

Remark 3: We may draw the conclusion from Theorem 2 that the improved correlation coefficient overcomes the limitations mentioned above of Xu’s correlation coefficients.

First of all, the improved correlation coefficient avoids emerging the 0/0 type quotient, which is more reasonable in mathematical logic.

In addition, under the condition A1A2, the improved correlation coefficient ρ*(A1, A2) is certainly less than one by the third property of Theorem 3, which is coincident with our intuition. For instance, in Example 1, A1A2 and ρ*(A1, A2)=0.9<1. Furthermore, the improved correlation coefficient can guarantee that the correlation coefficient of any two IFSs equals 1 if and only if these two IFSs are the same. Consequently, it can be applied in medical diagnosis and clustering effectively, which will be discussed in Section 4.

In many situations, the weight of every element xiX should be taken into account. For example, in the multiple attribute decision-making problems, each attribute usually has different importance and, thus, needs to be assigned a different weight. As a result, we further extend Formula (3).

Definition 9: Let A1, A2IFS(X), and X={x1, x2, …, xn} be a finite universe of discourse, then we define

(4)ρ1(A1,A2)=12i=1n{ωi[αi(1Δμi)+βi(1Δνi)]}

as a correlation coefficient of the IFSs A1 and A2, where

αi=cΔμiΔμmaxcΔμminΔμmax,  βi=cΔνiΔνmaxcΔνminΔνmax(c>2),

the weight vector of xi (i=1, 2, …, n) is ω=(ω1, ω2, …, ωn)T, ωi≥0<mspace;(i=1, 2, …, n) and i=1nωi=1. The value of ωi can be determined by several methods, such as statistical distribution, analytic hierarchy process, coefficient of variation method, and so on. It can also be determined according to the experts’ opinions. Especially, if ω=(1n,1n,,1n)T, Formula (4) reduces to Formula (3).

3.2 The Improved Correlation Coefficient of the IVIFSs

In the following, we generalize the idea of Definition 8 to the IVIFS theory.

Definition 10: Let Ã1, Ã2IVIFS(X), and X={x1, x2, …, xn} be a finite universe of discourse, then we define

(5)ρ2(A˜1,A˜2)=14ni=1n[γi(1Δμ˜iL)+ζi(1Δμ˜iU)+φi(1Δν˜iL)+ψi(1Δν˜iU)]

as a correlation coefficient of IVIFSs Ã1 and Ã2, where

γi=cΔμ˜iLΔμ˜maxLcΔμ˜minLΔμ˜maxL,   ζi=cΔμ˜iUΔμ˜maxUcΔμ˜minUΔμ˜maxU,φi=cΔν˜iLΔν˜maxLcΔν˜minLΔν˜maxL,   ψi=cΔν˜iUΔν˜maxUcΔν˜minUΔν˜maxU(c>2),Δμ˜iL=|μ˜A˜1L(xi)μ˜A˜2L(xi)|,   Δν˜iL=|ν˜A˜1L(xi)ν˜A˜2L(xi)|,Δμ˜iU=|μ˜A˜1U(xi)μ˜A˜2U(xi)|,   Δν˜iU=|ν˜A˜1U(xi)ν˜A˜2U(xi)|(i=1,2,,n),Δμ˜minL=mini{|μ˜A˜1L(xi)μ˜A˜2L(xi)|},Δμ˜minU=mini{|μ˜A˜1U(xi)μ˜A˜2U(xi)|},Δμ˜maxL=maxi{|μ˜A˜1L(xi)μ˜A˜2L(xi)|},Δμ˜maxU=maxi{|μ˜A˜1U(xi)μ˜A˜2U(xi)|},Δν˜minL=mini{|ν˜A˜1L(xi)ν˜A˜2L(xi)|},Δν˜minU=mini{|ν˜A˜1U(xi)ν˜A˜2U(xi)|},Δν˜maxL=maxi{|ν˜A˜1L(xi)ν˜A˜2L(xi)|},Δν˜maxU=maxi{|ν˜A˜1U(xi)ν˜A˜2U(xi)|}.

Equation (5) also can be generalized to more general cases:

(6)ρ3(A˜1,A˜2)=14i=1n{ωi[γi(1Δμ˜iL)+ζi(1Δμ˜iU)+φi(1Δν˜iL)+ψi(1Δν˜iU)]},

where ω=(ω1, ω2, …, ωn)T, ωi ≥ 0 (i=1, 2, …, n) is the weight of xi (i=1, 2, …, n) and i=1nωi=1. The value of ωi can be determined by the same way which was mentioned in Definition 9. Especially, if ω=(1n,1n,,1n)T, Formula (6) reduces to Formula (5).

Theorem 3:The correlation coefficient ρ2(A˜1,A˜2) satisfies the following properties:

  1. ρ2(A˜1,A˜2)=ρ2(A˜2,A˜1);

  2. 0ρ2(A˜1,A˜2)1;

  3. A˜1=A˜2ρ2(A˜1,A˜2)=1.

Proof

Obviously.

Since

0<γi<1, 0<ζ1, 0<φi1, 0<ψi1

and

01Δμ˜iL1,01Δμ˜iU1,01Δν˜iL1,01Δν˜iU1,

then

0γi(1Δμ˜iL)+ζi(1Δμ˜iU)+φi(1Δν˜iL)+ψi(1Δν˜iU)4

where i=1, 2, …, n. Thus, by Eq. (5), we know that 0ρ2(A˜1,A˜2)1.

“⇒” If Ã1=Ã2, it implies that: μ˜A˜1L(xi)=μ˜A˜2L(xi),μ˜A˜1U(xi)=μ˜A˜2U(xi),ν˜A˜1L(xi)=ν˜A˜2L(xi),ν˜A˜1U(xi)=ν˜A˜2U(xi) for all xiX, then

Δμ˜iL=Δμ˜minL=Δμ˜maxL=0,   Δμ˜iU=Δμ˜minU=Δμ˜maxU=0,Δν˜iL=Δν˜minL=Δν˜maxL=0,   Δν˜iU=Δν˜minU=Δν˜maxU=0.

Thus, ρ2(A˜1,A˜2)=1.

“⇐” If ρ2(A˜1,A˜2)=1, by the fact that

0γi(1Δμ˜iL)1,0ζi(1Δμ˜iU)1,0φi(1Δν˜iL)1,0ψi(1Δν˜iU)1,

we have

γi(1Δμ˜iL)+ζi(1Δμ˜iU)+φi(1Δν˜iL)+ψi(1Δν˜iU)=4

and

γi(1Δμ˜iL)=ζi(1Δμ˜iU)=φi(1Δν˜iL)=ψi(1Δν˜iU)=1.

Since

0<γi<1, 0<ζ1, 0<φi1, 0<ψi1

and

01Δμ˜iL1,   01Δμ˜iU1,   01Δν˜iL1,   01Δν˜iU1,

then, we obtain

1Δμ˜iL=1Δμ˜iU=1Δν˜iL=1Δν˜iU=1,

i.e.,

Δμ˜iL=Δμ˜iU=Δν˜iL=Δν˜iU=0.

Thus,

μ˜A˜1L(xi)=μ˜A˜2L(xi),   μ˜A˜1U(xi)=μ˜A˜2U(xi),ν˜A˜1L(xi)=ν˜A˜2L(xi),   ν˜A˜1U(xi)=ν˜A˜2U(xi),

for all xiX, i.e., Ã1=Ã2.

4 Applications of the Improved Correlation Coefficient

In this section, we mainly focus on the applications of the improved correlation coefficient in medical diagnosis and clustering.

4.1 The Application in Medical Diagnosis

First, for the set of diagnosis A={A1, A2, …, An}, the set of symptoms C={C1, C2, …, Ct}, and the set of patients B={B1, B2, …, Bm}, suppose the weight of Ck is ωk (k=1, 2, …, t), then the diagnosis steps are as follows:

  • Step 1: Based on the experience of experts, each symptom is described by a pair of parameters (μ, ν), i.e., the membership μ and the non-membership ν.

  • Step 2: For every patient Bj (j=1, 2, …, m), according to Formula (4), we can calculate the correlation coefficient ρ1(Ai,Bj),(i=1,2,,n), where Ai (i=1, 2, …, n) are the diagnosis results.

  • Step 3: For every patient Bj (j=1, 2, …, m), select the diagnosis result Ai0 in A most close to the patient Bj, i.e., ρ1(Ai0,Bj)=max{ρ1(Ai,Bj)|AiA}. Hence, it can be asserted that the diagnosis result of patient Bj is Ai0.

Example 3 ([14]): To make a proper diagnosis A for a patient with the given values of symptoms C, a medical knowledge base is necessary that involves elements described in terms of IFSs. The set of symptoms is

C={C1,C2,C3,C4,C5}={Temperature,Headache,Stomachpain,Cough,Chestpain},

the set of diagnosis is

A={A1,A2,A3,A4,A5}={Viralfever,Malaria,Typhoid,Stomachproblem,Chestproblem}

and the set of patients is

B={B1,B2,B3,B4}={Al,Bob,Joe,Ted}.

ω=(ω1, ω2, …, ω5)T is the weight vector of symptoms, let ω1=ω2==ω5=15. The data are given in Table 1 – each symptom is described by a pair of parameters (μ, ν), i.e., the membership μ and the non-membership ν. The symptoms are given in Table 2. We need to seek a diagnosis for each patient Bj (j=1, 2, 3, 4).

Table 1:

Symptoms Characteristic for the Considered Diagnoses.

A1A2A3A4A5
C1(0.4, 0.0)(0.7, 0.0)(0.3, 0.3)(0.1, 0.7)(0.1, 0.8)
C2(0.3, 0.5)(0.2, 0.6)(0.6, 0.1)(0.2, 0.4)(0.0, 0.8)
C3(0.1, 0.7)(0.0, 0.9)(0.2, 0.7)(0.8, 0.0)(0.2, 0.8)
C4(0.4, 0.3)(0.7, 0.0)(0.2, 0.6)(0.2, 0.7)(0.2, 0.8)
C5(0.1, 0.7)(0.1, 0.8)(0.1, 0.9)(0.2, 0.7)(0.8, 0.1)
Table 2:

Symptoms Characteristic for the Considered Patients.

C1C2C3C4C5
B1(0.8, 0.1)(0.6, 0.1)(0.2, 0.8)(0.6, 0.1)(0.1, 0.6)
B2(0.0, 0.8)(0.4, 0.4)(0.6, 0.1)(0.1, 0.7)(0.1, 0.8)
B3(0.8, 0.1)(0.8, 0.1)(0.0, 0.6)(0.2, 0.7)(0.0, 0.5)
B4(0.6, 0.1)(0.5, 0.4)(0.3, 0.4)(0.7, 0.2)(0.3, 0.4)

First, construct IFSs:

A1={C1,0.4,0.0,C2,0.3,0.5,C3,0.1,0.7,C4,0.4,0.3,C5,0.1,0.7},A2={C1,0.7,0.0,C2,0.2,0.6,C3,0.0,0.9,C4,0.7,0.0,C5,0.1,0.8},A3={C1,0.3,0.3,C2,0.6,0.1,C3,0.2,0.7,C4,0.2,0.6,C5,0.1,0.9},A4={C1,0.1,0.7,C2,0.2,0.4,C3,0.8,0.0,C4,0.2,0.7,C5,0.2,0.7},A5={C1,0.1,0.8,C2,0.0,0.8,C3,0.2,0.8,C4,0.2,0.8,C5,0.8,0.1},B1={C1,0.8,0.1,C2,0.6,0.1,C3,0.2,0.8,C4,0.6,0.1,C5,0.1,0.6},B2={C1,0.0,0.8,C2,0.4,0.4,C3,0.6,0.1,C4,0.1,0.7,C5,0.1,0.8},B3={C1,0.8,0.1,C2,0.8,0.1,C3,0.0,0.6,C4,0.2,0.7,C5,0.0,0.5},B4={C1,0.6,0.1,C2,0.5,0.4,C3,0.3,0.4,C4,0.7,0.2,C5,0.3,0.4}.

Then, we utilize the improved correlation coefficient to derive a diagnosis for each patient Bj (j=1, 2, 3, 4). Let c=3 in Formula (4). Then, all the diagnosis results for the considered patients are listed in Table 3.

Table 3:

Correlation Coefficients of Symptoms for Each Patient.

A1A2A3A4A5
B10.77270.78600.75570.47400.4697
B20.61360.47170.68960.88480.5828
B30.71320.62210.77130.52960.4805
B40.78640.72270.67470.56000.4987

Finally, from the arguments in Table 3, we derive a proper diagnosis as follows:

Al developed malaria; Bob had some problems with his stomach; Joe got typhoid; Ted got viral fever.

Based on the correlation coefficient (1), Xu’s diagnosis results [19] (Al suffered from malaria, Bob from a stomach problem, and both Joe and Ted from viral fever) are different from the results obtained by the improved correlation coefficient (4). Because the improved correlation coefficient has a wider scope of application and can solve the problems of the existing correlation coefficient (mentioned in Remark 2), the diagnosis results obtained by the improved correlation coefficient (4) are more reasonable. Furthermore, Vlachos and Sergiadis [16] made diagnosis utilizing the symmetric discrimination information measure DIFS(A, B) and pointed out that the diagnosis results of DIFS(A, B) were more effective than the results of distance-based methods and the similarity-dissimilarity measure. The diagnosis results obtained by DIFS(A, B) were as follows: Al got viral fever, Bob had some problems with his stomach, Joe got typhoid, and Ted got viral fever, which are also different from those by the improved correlation coefficient (4). The differences are because that the results derived using DIFS(A, B) are prone to the influence of unfair arguments with too high or too low values, while the improved correlation coefficient (4) can relieve the influence of these unfair arguments by emphasizing the role of the considered arguments as a whole [21].

4.2 The Application in Clustering

In the following, we will show that the improved correlation coefficient (4) is more effective than correlation coefficients (1) and (2) in clustering.

For the object sets Ai (i=1, 2, …, m), suppose that the weight vector of xi (i=1, 2, …, n) is ω=(ω1, ω2, …, ωn)T, then we discuss the clustering steps:

  • Step 1: Utilize Eq. (4) to calculate the correlation coefficient ρ1(Ai,Aj) between object sets Ai and Aj (i, j=1, 2, …, m), and construct association matrix C˙=(ρ˙ij)m×m, where ρ˙ij=ρ1(Ai,Aj)(i,j=1,2,,m).

  • Step 2: Obtain equivalence association matrix C˙¯=(ρ˙¯ij)m×m by finite times compositions of C˙=(ρ˙ij)m×m and construct λ-cutting matrix C˙¯λ=(λρ¯ij)m×n of the equivalence association matrix C˙¯.

  • Step 3: For the λ-cutting matrix C˙¯λ, if all the elements of the ith row or column are the same with the jth row or column, we assert object sets Ai and Aj are in the same class.

Example 4 ([28]): A car market is going to classify five different cars. Every car has six evaluation factors: (1) G1 − fuel consumption; (2) G2 − degree of friction; (3 )G3 − the price; (4) G4 − degree of comfort; (5) G5 − design; (6) G6 − security. The information of every car under each evaluation factor is represented by intuitionistic fuzzy numbers, which are shown in Table 4.

Table 4:

Information.

x1x1x3x4x5x6
A1(0.3, 0.5)(0.6, 0.1)(0.4, 0.3)(0.8, 0.1)(0.1, 0.6)(0.5, 0.4)
A2(0.6, 0.3)(0.5, 0.2)(0.6, 0.1)(0.7, 0.1)(0.3, 0.6)(0.4, 0.3)
A3(0.4, 0.4)(0.8, 0.1)(0.5, 0.1)(0.6, 0.2)(0.4, 0.5)(0.3, 0.2)
A4(0.2, 0.4)(0.4, 0.1)(0.9, 0.0)(0.8, 0.1)(0.2, 0.5)(0.7, 0.1)
A5(0.5, 0.2)(0.3, 0.6)(0.6, 0.3)(0.7, 0.1)(0.6, 0.2)(0.5, 0.3)

If the weight vector of the element xi (i=1, 2, …, 6) is ω=(16,16,,16)T, let c=3, then we can classify the cars by the improved correlation coefficient ρ1(Ai,Aj).

First, we utilize the improved correlation coefficient ρ1(Ai,Aj) to calculate the correlation between each pair of IFSs Ai and Aj (i, j=1, 2, …, 5), and construct an association matrix:

C=[10.84220.82030.79810.68850.842210.86340.81090.80960.82030.863410.80470.75350.79810.81090.804710.70730.68850.80960.75350.70731]

Then, we obtain an equivalent association matrix by finite times compositions of C. For

C2=[10.84220.84220.81090.80960.842210.86340.81090.80960.84220.863410.81090.80960.81090.81090.810910.80960.80960.80960.80960.80961]C

C4=[10.84220.84220.81090.80960.842210.86340.81090.80960.84220.863410.81090.80960.81090.81090.810910.80960.80960.80960.80960.80961]=C2,

we get the equivalent association matrix C2, denoted by C˜.

Finally, due to the fact that the confidence level λ has a close relationship with the elements of the equivalent association matrix C˜, we give a detailed sensitivity analysis with respect to the confidence level λ and get all the possible classifications of the five cars Ai (i=1, 2, 3, 4, 5).

  1. if 0λ0.8096,C˜λ=[1111111111111111111111111]

    then the cars are of the same type: {A1, A2, …, A5};

  2. if 0.8096<λ0.8109,C˜λ=[1111011110111101111000001]

    then the cars are classified into the following two types: {A1, A2, A3, A4}, {A5};

  3. if 0.8109<λ0.8422,C˜λ=[1110011100111000001000001]

    then the cars are classified into the following three types: {A1, A2, A3}, {A4}, {A5};

  4. if 0.8422<λ0.8634,C˜λ=[1000001100011000001000001]

    then the cars are classified into the following four types: {A1}, {A2, A3}, {A4}, {A5};

  5. if 0.8634<λ1,C˜λ=[1000001000001000001000001]

    then the cars are classified into the following five types: {A1}, {A2}, {A3}, {A4}, {A5}.

From the above analysis, we know that the improved correlation coefficient (4) can be applied in the clustering of IFSs effectively. Moreover, there are five situations based on the association matrix, which is constructed by the improved correlation coefficient (4), while only three situations were obtained in Ref. [28]. It can be seen that the improved correlation coefficient has higher accuracy in clustering.

5 Conclusion

The existing correlation coefficients did not meet some desirable properties in the IFS theory. In order to solve this problem, we have introduced an improved correlation coefficient of the IFSs. We have shown the advantages of the improved correlation coefficient by comparing with the correlation coefficients proposed by Xu [19, 21]. Then, we have further generalized it to the IVIFS environment. Finally, the improved correlation coefficient has also been applied effectively in medical diagnosis and clustering.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (11571158).

Bibliography

[1] K. Atanassov, Intuitionistic fuzzy sets, Fuzzy Sets Syst.20 (1986), 87–96.10.1016/S0165-0114(86)80034-3Search in Google Scholar

[2] K. Atanassov and G. Gargov, Interval-valued intuitionistic fuzzy sets, Fuzzy Sets Syst.31 (1989), 343–349.10.1016/0165-0114(89)90205-4Search in Google Scholar

[3] N. Chen, Z. S. Xu and M. M. Xia, Correlation coefficients of hesitant fuzzy sets and their applications to clustering analysis, Appl. Math. Model.37 (2013), 2197–2211.10.1016/j.apm.2012.04.031Search in Google Scholar

[4] C. H. Chu, K. C. Hung and P. Julian, A complete pattern recognition approach under Atanassov’s intuitionistic fuzzy sets, Knowl. Based Syst.66 (2014), 36–45.10.1016/j.knosys.2014.04.014Search in Google Scholar

[5] G. Deschrijver and E. E. Kerre, On the relationship between some extensions of fuzzy set theory, Fuzzy Sets Syst.133 (2003), 227–235.10.1016/S0165-0114(02)00127-6Search in Google Scholar

[6] A. Fiche, J. Cexus, A. Martin and A. Khenchaf, Features modeling with an α-stable distribution: application to pattern recognition based on continuous belief functions, Inf. Fusion14 (2013), 504–520.10.1016/j.inffus.2013.02.004Search in Google Scholar

[7] J. Goguen, L-fuzzy sets, J. Math. Anal. Appl.18 (1967), 145–174.10.1016/0022-247X(67)90189-8Search in Google Scholar

[8] C. Grosan, A. Abraham and S. Tigan, Multicriteria programming in medical diagnosis and treatments, Appl. Soft Comput.8 (2008), 1407–1417.10.1016/j.asoc.2007.10.014Search in Google Scholar

[9] H. L. Huang, (T,S)-Based interval-valued intuitionistic fuzzy composition matrix and its application for clustering, Iran. J. Fuzzy Syst.9 (2012), 7–19.Search in Google Scholar

[10] W. L. Hung, Using statistical viewpoint in developing correlation of intuitionistic fuzzy sets, Int. J. Uncertain. Fuzziness Knowl. Based Syst.9 (2001), 509–516.10.1142/S0218488501000910Search in Google Scholar

[11] W. L. Hung and J. W. Wu, Correlation of intuitionistic fuzzy sets by centroid method, Inf. Sci.144 (2002), 219–225.10.1016/S0020-0255(02)00181-0Search in Google Scholar

[12] H. Laanaya, A. Martin, D. Aboutajdine and A. Khenchaf, Support vector regression of membership functions and belief functions – application for pattern recognition, Inf. Fusion11 (2010), 338–350.10.1016/j.inffus.2009.12.007Search in Google Scholar

[13] M. Z. Ranitović, Lattice representations of interval-valued fuzzy sets, Fuzzy Sets Syst.236 (2014), 50–57.10.1016/j.fss.2013.07.006Search in Google Scholar

[14] E. Szmidt and J. Kacprzyk, A similarity measure for intuitionistic fuzzy sets and its application in supporting medical diagnostic reasoning, Lect. Notes Artif. Intell.3070 (2004), 388–393.10.1007/978-3-540-24844-6_56Search in Google Scholar

[15] Z. Takáč, Inclusion and subsethood measure for interval-valued fuzzy sets and for continuous type-2 fuzzy sets, Fuzzy Sets Syst.224 (2013), 106–120.10.1016/j.fss.2013.01.002Search in Google Scholar

[16] I. K. Vlachos and D. D. Sergiadis, Intuitionistic fuzzy information-application to pattern recognition, Pattern Recognit. Lett.28 (2007), 197–206.10.1016/j.patrec.2006.07.004Search in Google Scholar

[17] P. Z. Wang, Fuzzy set theory and applications, Shanghai Scientific and Technical Publishers, Shanghai, 1983.Search in Google Scholar

[18] G. J. Wang and Y. Y. He, Intuitionistic fuzzy sets and L-fuzzy sets, Fuzzy Sets Syst.110 (2000), 271–274.10.1016/S0165-0114(98)00011-6Search in Google Scholar

[19] Z. S. Xu, On correlation measures of intuitionistic fuzzy sets, Lect. Notes Comput. Sci.4224 (2006), 16–24.10.1007/11875581_2Search in Google Scholar

[20] Z. S. Xu, Hesitant fuzzy multi-attribute decision making based on TOPSIS with incomplete weight information, Knowl. Based Syst.52 (2013), 53–64.10.1016/j.knosys.2013.05.011Search in Google Scholar

[21] Z. S. Xu and X. Q. Cai, Intuitionistic fuzzy information aggregation: theory and applications, Springer, Heidelberg, 2012.10.1007/978-3-642-29584-3Search in Google Scholar

[22] Z. S. Xu, J. Chen and J. J. Wu, Clustering algorithm for intuitionistic fuzzy sets, Inf. Sci.19 (2008), 3775–3790.10.1016/j.ins.2008.06.008Search in Google Scholar

[23] W. Yao, On many-valued stratified L-fuzzy convergence spaces, Fuzzy Sets Syst.159 (2008), 2503–2519.10.1016/j.fss.2008.03.003Search in Google Scholar

[24] Z. L. Yue, Group decision making with multi-attribute interval data, Inf. Fusion14 (2013), 551–561.10.1016/j.inffus.2013.01.003Search in Google Scholar

[25] S. H. Yue, T. Wu, J. Pan and H. Wang, Fuzzy clustering based ET image fusion, Inf. Fusion14 (2013), 487–497.10.1016/j.inffus.2012.09.004Search in Google Scholar

[26] L. A. Zadeh, Fuzzy sets, Inf. Control8 (1965), 338–353.10.1016/S0019-9958(65)90241-XSearch in Google Scholar

[27] L. A. Zadeh, Outline of a new approach to the analyses of complex systems and decision processes interval-valued fuzzy sets, IEEE Trans. Syst. Man Cybern.3 (1973), 28–44.10.1109/TSMC.1973.5408575Search in Google Scholar

[28] H. M. Zhang, Z. S. Xu and Q. Chen, Clustering method of intuitionistic fuzzy sets, Control Decis.22 (2007), 882–888 (in Chinese).Search in Google Scholar

[29] H. Y. Zhang, W. X. Zhang and C. L. Mei, Entropy of interval-valued fuzzy sets based on distance and its relationship with similarity measure, Knowl. Based Syst.22 (2009), 449–454.10.1016/j.knosys.2009.06.007Search in Google Scholar

[30] Z. H. Zhang, J. Y. Yang, Y. P. Ye and Q. Zhang, A type of score function on intuitionistic fuzzy sets with double parameters and its application to pattern recognition and medical diagnosis, Procedia Eng.29 (2012), 4336–4342.10.1016/j.proeng.2012.01.667Search in Google Scholar

Received: 2017-03-12
Published Online: 2017-07-06
Published in Print: 2019-04-24

©2019 Walter de Gruyter GmbH, Berlin/Boston

This article is distributed under the terms of the Creative Commons Attribution Non-Commercial License, which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Downloaded on 7.5.2024 from https://www.degruyter.com/document/doi/10.1515/jisys-2017-0094/html
Scroll to top button