• Tidak ada hasil yang ditemukan

Feature-reduction fuzzy co-clustering approach for hyper-spectral image analysis

N/A
N/A
Nguyễn Gia Hào

Academic year: 2023

Membagikan "Feature-reduction fuzzy co-clustering approach for hyper-spectral image analysis"

Copied!
15
0
0

Teks penuh

(1)

Knowledge-Based Systems 216 (2021) 106549

Contents lists available atScienceDirect

Knowledge-Based Systems

journal homepage:www.elsevier.com/locate/knosys

Feature-reduction fuzzy co-clustering approach for hyper-spectral image analysis

Nha Van Pham

a,b

, Long The Pham

a

, Witold Pedrycz

c

, Long Thanh Ngo

a,

aDepartment of Information Systems, Le Quy Don Technical University, Hanoi, Viet Nam

bMIST Institute of Science and Technology, 17 Hoang Sam, Hanoi, Viet Nam

cDepartment of Electrical & Computer Engineering, University of Alberta, Edmonton T6R 2V4 AB, Canada

a r t i c l e i n f o

Article history:

Received 1 January 2020

Received in revised form 2 August 2020 Accepted 19 October 2020

Available online 26 December 2020 Keywords:

Fuzzy co-clustering Dimensionality reduction Cluster tendency

Hyper-spectral satellite image Land-cover classification

a b s t r a c t

Fuzzy co-clustering algorithms are the effective techniques for multi-dimensional clustering in which all features are considered of equal importance (relevance). In fact, the features’ importance could be different, even several of them could be considered redundant. The removal of the redundant features has formed the idea of feature-reduction in problems of the big data processing. In this paper, we propose a new unsupervised learning scheme by incorporating the feature-weighted entropy into the objective function of fuzzy co-clustering, called the Feature-Reduction Fuzzy Co-Clustering Algorithm (FRFCoC). First, a new objective function is formed on the basis of the original fuzzy co-clustering objective function which adds parameters representing the entropy weight of the different features.

Next, a feature-reduction and clustering automatic schema are adjusted based on FCoC’s original learning schema which calculates new parameters and conditions to eliminate irrelevant feature components. FRFCoC algorithm can be mathematically shown to converge after a finite number of iterations. The experiment results were conducted on some many-features data sets and hyperspectral images that have demonstrated the outstanding performance of FRFCoC algorithm compared with some previously proposed algorithms.

©2020 Elsevier B.V. All rights reserved.

1. Introduction

The hyperspectral imaging is a progressive development tech- nology adhering the development of aeronautics and space re- mote sensing. It combines the two latest technologies in imagery and spectroscopy. Through the hyperspectral imaging sensor, se- quential imagery data can be acquired in narrow bands with a high spectral resolution which benefits the theoretical research on hyperspectral data analysis in various fields [1]. Today, this technology is widely used in different fields, such as military, search salvage and rescue, environmental monitoring, mineral exploration and public security. The advantages of hyperspec- tral imaging are the high spectral resolution, providing spec- tral characteristics and spatial information simultaneously. The large number of bands contains knowledge for target detec- tion and recognition. In fact, the hyperspectral images contain many redundant feature components, reducing the performance and quality in hyperspectral image processing [2,3] . Therefore, the dimensionality reduction is an important phase to enhance

Corresponding author.

E-mail addresses: famvannha@gmail.com(N.V. Pham),longpt@mta.edu.vn (L.T. Pham),wpedrycz@ualberta.ca(W. Pedrycz),ngotlong@mta.edu.vn (L.T. Ngo).

the performance. Some methods of dimensionality reduction are previously mentioned as band selection [4,5], discriminant anal- ysis [6–8], principal component analysis [9–11]. Recently, Yang et al. have proposed an algorithm called a feature-reduction FCM (FRFCM) [12] that can automatically compute feature weights by adding feature-weighted entropy into the FCM objective function.

The algorithm exhibits a feature-reduction schema to eliminate irrelevant features taking the lowest weights to decrease the computational time and improve clustering performance.

The co-clustering is a useful tool for data analysis, where data objects are grouped into a number of clusters according to their similarity in multi-dimensional spaces. These techniques can si- multaneously perform on both spatial dimensions and feature dimension, which are suitable with problems of complex data like the multi-dimensional data [13,14] or hyperspectral image.

Some co-clustering techniques have been studied and improved to resolve some problems such as clustering documents and keywords [15–17], color segmentation [18], categorical multi- variate data [15] and high-dimensional data [19]. Recently, we have proposed a co-clustering algorithm by combining advan- tages of fuzzy co-clustering and interval-valued fuzzy sets. We called IVFCoC algorithm [20]. IVFCoC algorithm was guided by the new objective function with two values of the fuzzifier pro- ducing the FOU. Experiments were conducted on the datasets

https://doi.org/10.1016/j.knosys.2020.106549 0950-7051/©2020 Elsevier B.V. All rights reserved.

(2)

N.V. Pham, L.T. Pham, W. Pedrycz et al. Knowledge-Based Systems 216 (2021) 106549

of color images, multi-spectral satellite images and high dimen- sional datasets and IVFCoC algorithm has proven to achieve the better performance than some previously proposed clustering algorithms. However, a limitation of IVFCoC algorithm is the high calculation complexity so when dealing with the data types of high-dimensional (number of features is large) such as hy- perspectral images, we have to pay an expensive price for the consumption time and the accuracy of clusters. In addition, co- clustering algorithms generally treat data points with feature components under equal importance. However, in fact there are various data sets with irrelevant features involved in clustering process that may cause bad performance for clustering algo- rithms. That is, different feature components should be weighted differently to reflect their relative importance. Then removing the lowest significant features will increase the performance of clustering [12].

Some variants of FCM algorithm had been proposed, such as weighted K-Means [21,22], weighted FCM uses feature-weight learning (WFCM) [23–27]. These works have been proposed by considering the various features of data by weighting these fea- tures in the scheme of algorithms to improve the accuracy of clusters. However, the feature-reduction schema has not been proposed in these works. This may be the motivation for Yang et al. [12] to propose a new schema to improve FCM by adding feature reduction by using feature-weighted entropy (FRFCM).

FRFCM algorithm can automatically compute individual feature weight, and simultaneously reduce these irrelevant feature com- ponents. Authors first consider the FCM objective function with feature-weighted entropy, and construct a learning schema for parameters, and then reduce these irrelevant feature compo- nents. A new procedure for eliminating irrelevant features with small weights is created for feature reduction. Experiments are conducted on some numerical and real data sets to compare FRFCM with various feature-weighted FCM methods in the litera- ture. Experimental results and comparisons actually demonstrate these good aspects of FRFCM with its effectiveness and usefulness in practice.

Currently, there are many clustering algorithms with different mathematical models and objective functions such as K-Means, Fuzzy C-Means, DBSCAN, etc [28]. However, most of these al- gorithms only consider data in terms of data objects. Feature components are only considered indirectly through data objects.

This means that in the objective functions of clustering algo- rithms there is no presence of the number of features parameter or the weight of the feature. Therefore, the feature-reduction for these algorithms is not mathematically significant though being significant in data pre-processing which is similar to band selection methods in the hyperspectral image [4,5]. Meanwhile, co-clustering algorithms in general, fuzzy co-clustering algorithm in particular always exist parameters of feature components such as the number of features and the feature weight and the feature membership function (as will be as described in Section 2 be- low). Therefore, co-clustering algorithms are more suitable than other clustering algorithms in the clustering of multi-dimensional and many features data types including the hyperspectral im- age [29]. We are motivated to create a new feature-reduction mathematical model that is specific to clustering algorithms.

In this paper, we propose a new unsupervised learning scheme by incorporating feature-weighted entropy in the objective func- tion of fuzzy co-clustering, called Feature-Reduction Fuzzy Co- Clustering algorithm (FRFCoC). First, a new objective function is formed on the basis of the original fuzzy co-clustering objec- tive function which adds parameters representing the entropy weight of the different features in the data. Next, a feature- reduction and clustering automatic schema was adjusted based on FCoC’s original learning schema which adds steps to cal- culate new parameters and conditions to eliminate irrelevant

feature components. Simultaneously, the authors have proved that FRFCoC algorithm can converge after a limited number of iterations using mathematical theorems and lemma. The exper- iments were conducted on some many-features data sets and hyperspectral images. Experimental results have shown that the outstanding performance of FRFCoC algorithm is compared with some previous proposed algorithms.

The rest of the paper is organized as follows. Section2briefly reviews some related works consisting of fuzzy co-clustering, feature-reduction based FCM; Section 3 presents the proposed framework of fuzzy co-clustering algorithm; Section 4 shows comparisons of the proposed method with other existing algo- rithms using validity indices; Section5includes conclusion and some future works.

2. Related works

In this section, we briefly presents fuzzy co-clustering algo- rithm and feature reduction based FCM algorithm.

2.1. Feature-reduction fuzzy clustering algorithm

The fuzzy clustering algorithm is also known as the Fuzzy C- Means (FCM) [30] algorithm which initially did not consider the feature components. That is, the objective function of FCM does not include the feature components of the data. Some the feature- reduction methods have been proposed to improve the perfor- mance of FCM [23–27]. However, data may contains irrelevant features therefore the feature-reduction becomes a necessary phase in FCM algorithm, especially for the many feature data pro- cessing. Yang et al. [12] proposed a new schema to improve FCM by adding feature reduction by using feature-weighted entropy (FRFCM). In this schema, each feature was assigned a weight that will be updated through iteration. Afterwards, features taking the smallest weight value will be eliminated after iterating.

LetNis the number of data objects;Cis the number of clusters (2

C

N); LetX

= (

x1

,

x2

, . . . ,

xN

)

, xi

X

,

i

=

1

, . . . ,

N, be the input dataset inD-dimensional feature space;xij

R

,

j

=

1

, . . . ,

Ddenotes featurejof data objecti;P

= (

p1

,

p2

, . . . ,

pC

)

, pc

P(c

=

1

, . . . ,

C)

,

pcj

R(j

=

1

, . . . ,

D) set of feature-based centroids inD-dimensional feature space. Denoteuci stands for the object membership degree of xi to cluster c, U

= {

uci

}

be the object membership degree matrix. The objective function of FRFCM algorithm is indicated in(1).

JFRFCM(U

,

V

,

W)

=

C

c=1 N

i=1 D

j=1

umci

δ

j

w

j(xij

pcj)2

+

Tw

D

j=1

(

w

jlog

δ

j

w

j) (1)

Where, W

= [ w

j

]

1xD with

w

j is the feature weight of the jth feature,

δ = [ δ

j

]

1xD with

δ

jis used to adjust feature weight

w

j. Tw is a coefficient calculated as follows:Tw

=

N

/

C [12]. That is, Tw depends on the size of dataNand the number of clusters of dataC.

The components of objective functionJFRFCM are defined by formulas(2),(3)and(4).

uci

=

( ∑

D j=1

δ

j

w

j

(

xij

pcj

)

2

)

1/(m1)

C q=1

( ∑

D

k=1

δ

k

w

k

(

xik

pck

)

2

)

1/(m1) (2)

w

j

=

δ1jexp

(

T1

w

C c=1

N

i=1umci

δ

j(xij

pcj)2

)

D k=1

δ1kexp

(

1

Tw

C c=1

N

i=1umci

δ

k(xik

pck)2

)

(3)

2

(3)

N.V. Pham, L.T. Pham, W. Pedrycz et al. Knowledge-Based Systems 216 (2021) 106549

pcj

=

N i=1ucixij

N

i=1uci (4)

There are two terms in the FRFCM objective function in Eq.(3).

The first term,

C c=1

N i=1

D j=1umci

δ

j

w

j

(

xij

pcj

)

2

is the sum of feature-weighted distances between data points and cluster cen- ters, which is minimized when the distance between points and centers is small. The second term, Tw

D

j=1

w

jlog

( δ

j

w

j

)

is a variant of the feature-weight entropy. Yang et al. [12] have explained why we use the constantTw

=

N

/

C to handle effects of the term

D

j=1

w

jlog

( δ

j

w

j

)

, which is also used in Eq.(3)as an updating equation for feature weights. As seen in Eq.(3), if the term

C

c=1

N i=1umci

δ

j

(

xij

pcj

)

2

is too large, then the numera- tor exp

(

− ∑

C c=1

N i=1umci

δ

j

(

xij

pcj

)

2

)

will become too small as close to a zero value. We need to avoid this case for preventing too many feature weights to be discarded during this updating step. On the other hand, if the term

C

c=1

N i=1umci

δ

j

(

xij

pcj

)

2

is too small, then the numerator exp

(

− ∑

C c=1

N i=1umci

δ

j

(

xij

pcj

)

2

)

will be closed to one so that it is difficult for the feature(s) to be discarded during the updating step. We also need to avoid this case. In this sense, we need to put a suitable constant to control it. In the FRFCM clustering algorithm, one goal is to cluster a data set (with N data points) intoC clusters. The numbersN and C are the two commonly given constants. We can use the constant Tw

=

N

/

C to control the term

C

c=1

N i=1umci

δ

j

(

xij

pcj

)

2

However, in the experiments of this paper, the formulaTw

=

N

/

C does not give exact results for both FRFCM and FRFCoC.

Therefore, in this paper, we consider Tw as a fuzzy parameter (fuzzy feature weight).

δ

j in

C c=1

N i=1

D

j=1umci

δ

j

w

j

(

xij

pcj

)

2

are used for control- ling the variants of feature weights, the choice of

δ

jis important.

To estimate the value of

δ

j, Yang et al. [12] have proposed a learning procedure. In probability theory and statistics, variance- to-mean ratio (VMR), like the coefficient of variation, is a nor- malized measure of the dispersion of a probability distribution.

VMR is a measure used to quantify whether a set of observed occurrences are clustered or dispersed compared to a standard statistical model. Smaller dispersion means the data set would be closer to the cluster center, while larger dispersion means the data set is far from the cluster center. VMR is defined as the ratio of the variance

σ

2 to the mean

µ

, VMR

= σ

2

. . Because we need to retain features that have small dispersion and then discard those which have large dispersion, we consider the reciprocal of VMR, i.e., mean-to-variance ratio (MVR). The coefficient of variation (CV) is defined as the ratio of the standard deviation

σ

to the mean, MVR

= σ/µ

[31]. Thus VMR and MVR can be used to calculate

δ

j[32]. However, [12] has proved experimentally that VMR and MVR cannot produce small weights.

And the term (Mean(x)j/Var(x)j) for the feature j can actually handle the dispersion between clusters in the data set. Therefore, we use (Mean(x)j

/

Var(x)j) to estimate

δ

j. That is, we consider the estimate for

δ

jas follows:

δ

j

=

(

Mean(x)j Var(x)j

)

(5) In which,Mean(x)jis the mean ofNdata points, andVar(x)jis the variance ofNdata points according tojthfeature.

To create a feature-reduction schema in the FRFCM algorithm, we need to select these unimportant features (i.e., small weights) during clustering processes. In our construction, we use a thresh- old to determine which features will be selected and then dis- carded.

The FRFCM algorithm diagram consists of the learning pro- cesses of membership function matrix U and feature weight matrixW that are shown as Algorithm1.

Algorithm 1FRFCM algorithm Input: Data set X

= {

xi

|

xi

RK

}

,

i

=

1

, ...,

N, the number of clustersC.

Output: The results of the clustering and feature reduction:

Centroids and the distribution of data points in each cluster and the number of reduction features.

1.Initialize the maximum error

ε

and the maximum number of iterations

τ

max.

2.Initializeucisuch that 0

uci

1 and the constraint conditions in Eq.(8).

3.Initialize randomly feature weight matrixW(0) 4.Set iteration

τ

=1.

5. REPEAT

6. Calculate and update

δ

jusing Eq.(5).

7. Calculate and updateuciusing Eq.(2).

8. Calculate and updatepcjusing Eq.(4).

9. Update

w

jusing Eq.(3).

10. Eliminate Dr features satisfying their weights are lower than the threshold, computeD(new)

=

D

Dr and setD

=

D(new). 11.

τ

=

τ

+1.

12. UNTIL

W(t)

W(t1)

 < ε

or

τ ≤ τ

max

The FRFCM algorithm is divided into three parts: (1) Compute the membership function,uci; (2) Update cluster centroid,pcj; and (3) Update the weight

w

j, which the computational complexity is O(

τ

NC2D), O(

τ

NC) and O(

τ

NCD2), respectively. The total one isO(

τ

NC2D

+ τ

NCD2). BecauseC is normally much smaller than D, therefore, the final computational complexity of FRFCM is O(

τ

NCD2). WithC is the number of clusters which is a known input parameter for each data set,Nis the number of data objects andDis the number of features of data.

Theoretically, FRFCM algorithm has proposed a new objec- tive function for fuzzy clustering based on the feature-weighted entropy and feature reduction techniques. Compared to fuzzy clustering (for example FCM), FRFCM considered data in detail for each feature through their weights and integrated a feature reduction scheme to eliminate small-weighted features. Some comparisons and experiments conducted on the synthetic and real data sets have demonstrated the superior performance of FRFCM compared to some traditional clustering algorithms. In this paper, we will use FRFCM’s contributions to improve fuzzy co-clustering in the many-feature data and the hyperspectral image processing.

2.2. Fuzzy co-clustering algorithm

Fuzzy co-clustering is a clustering method which follows the same principle as general co-clustering. They are simultaneously perform to clustering objects and features [33]. The difference is that the boundary between any two clusters is described in terms of membership functions rather than characteristic func- tions. Each work in the literature has applied fuzzy co-clustering to describe and solve specific problems in their own way. The authors in [34] denote fuzzy clustering algorithm as FCCM for cat- egorical multivariate data; the fuzzy clustering algorithm Fuzzy CoDoK in [15] for clustering documents and keywords; the fuzzy co-clustering algorithm FCCI is described in [18] for the color image segmentation problem class and the fuzzy co-clustering algorithm ibFCC described in [35] for the biomedical data seg- mentation. Recently, in [20] we denoted a fuzzy co-clustering algorithm as FCoC to propose IVFCoC algorithm by combining advantages of fuzzy co-clustering and interval-valued fuzzy sets which was aimed at solving clustering problems in the presence

(4)

N.V. Pham, L.T. Pham, W. Pedrycz et al. Knowledge-Based Systems 216 (2021) 106549

of complex data. These fuzzy co-clustering algorithms share the same principle where co-clustering is seen as a combination of object partitioning and feature ranking. Although the symbols of fuzzy co-clustering algorithms on different problems are differ- ent, their main contents are summarized in a uniform manner as follows.

Let N is the number of data objects; C is the number of clusters (2

C

N); Tu and Tv are fuzzy weights; LetX

= (

x1

,

x2

, . . . ,

xN

)

, xi

X

,

i

=

1

, . . . ,

N, be the input dataset inD- dimensional feature space;xij

R

,

j

=

1

, . . . ,

Ddenotes featurej of data objecti;P

= (

p1

,

p2

, . . . ,

pC

)

,pc

P(c

=

1

, . . . ,

C)

,

pcj

R(j

=

1

, . . . ,

D) set of feature-based centroids inD-dimensional feature space.

Denote uci stands for the object membership degree of xito cluster c,U

= {

uci

}

be the object membership degree matrix;

v

cj stands for the feature membership degree defined as the membership grade of featurejto the clustercandV

= {

v

cj

}

be the feature membership matrix with sizeCxD. Letdcijbe distance betweenxijandpcjgiven by:

dcij

= 

xij

,

pcj

2

=

(xij

pcj)2 (6)

The objective functionJFCoC(U

,

V

,

P) is revised in Eq.(7).

JFCoC(U

,

V

,

P)

=

C

c=1 N

i=1 D

j=1

uci

v

cjdcij

+

TU

C

c=1 N

i=1

uciloguci

+

TV

C

c=1 D

j=1

v

cjlog

v

cj (7)

To get optimal clustering results, the objective function FCoC (7) is minimized subject by the object membership func- tionsuciand the feature membership functions

v

cjare specific to each cluster to the following constraints:

C

c=1

uci

=

1

,

uci

[0

,

1]

, ∀

i

=

1

,

N

, ∀

c

=

1

,

C

D

j=1

v

cj

=

1

, v

cj

[0

,

1]

, ∀

j

=

1

,

D

, ∀

c

=

1

,

C

(8)

That is, every data object (

i

=

1

,

N)), its membership degree on each cluster receives a value in the range [0,1] (uci

[0

,

1], uci

[0

,

1]), and the total membership degree of a data object on all clusters is 1.0 (

C

c=1uci

=

1). And every data feature (

j

=

1

,

D), its membership degree on each cluster receives a value in the range [0,1] (

v

cj

[0

,

1],

c

=

1

,

C)), and the total membership degree of a data feature on all clusters is 1.0 (

C

c=1

v

cj

=

1) The components of objective functionJFCoC are defined by the formulas(9),(10)and(11).

uci

=

e

(D j=1

vcj dcij Tu )

C q=1e(

D j=1

vqj dqij Tu )

(9)

v

cj

=

e

(N i=1

ucidcij Tv )

D k=1e(

N i=1ucidcik

Tu )

(10)

pcj

=

N i=1ucixij

N

i=1uci (11)

The FCoC algorithm are shown in Algorithm 2. The computa- tional complexity of the FCoC algorithm isO(CND). Where,C is the number of clusters which is a known input parameter for each data set,Nis the number of data objects andDis the number of features of data.

Algorithm 2FCoC algorithm Input: Data set X

= {

xi

|

xi

RD

}

,

i

=

1

, ...,

N, the number of clusters isC.

Output:The clustering results: Centroids and the distribution of data points in each cluster.

1.Initialize the parameters Tu, Tv, maximum error

ε

and the maximum number of iterations

τ

max.

2.Initializeucisuch that 0

uci

1 and the constraint conditions in Eq.(8).

3.Set iteration number

τ

=1.

4.REPEAT

5. Calculate and updatepcjusing Eq.(11).

6. Calculate and updatedcijusing Eq.(6).

7. Calculate and update

v

cjusing Eq.(10).

8. Calculate and updateuciusing Eq.(9).

9. Increase

τ

=

τ

+1.

10.UNTIL(max(

|

uci

[ τ ] −

uci

[ τ −

1

]|

)

≤ ε

or

τ > τ

max).

The difference in mathematical model between FCoC and tra- ditional fuzzy clustering algorithms (for example FCM) is their objective functions. The objective function of FCM only considers data according to data object terminology. The objective function of FCoC considers data according to data object terminology and characteristics of data. That is, FCM only considers the number of data objects. Meanwhile, FCoC considers each feature weight of data objects. Therefore, FCoC can be considered as feature- weighted fuzzy clustering which can be used instead of FCM in the multi-dimensional and many features processing. Recent experiments of FCoC on the color image [18] and the biomed- ical image [35], documents [15], the many-featured data [20, 34] and multi-spectral images and hyperspectral images [20, 29] have shown that the performance of FCoC is better than traditional fuzzy cluster. However, the consideration of feature components leads to the problem of quantifying the importance of features that increase the computational complexity. Therefore, the feature-reduction can be a suitable method to improve the efficiency of FCoC. In this paper, we will use the feature-weighted and feature-reduction entropy techniques in [12] to build a new feature-reduction model for FCoC.

3. Feature-Reduction Fuzzy Co-Clustering algorithm

In this section, we will present a new FRFCoC clustering algo- rithm and prove mathematically convergence of FRFCoC.

3.1. FRFCoC algorithm

In this section, we present a new method to improve the performance of the algorithm FCoC by applying feature reduction, which called FRFCoC algorithm. In this method, each feature is assigned a weight, which is adjusted in iterations. Then, features taking the smallest weights will be removed. Considering theD- dimensional datasetX

= {

x1

,

x2

, . . . ,

xN

}

, the weights of features are W

= { w

1

, w

2

, . . . , w

D

}

. The objective function of FRFCoC algorithm is modified as follows,

JFRFCoC(U

,

V

,

P

,

W

, δ

)

=

C

c=1 N

i=1 D

j=1

uci

v

cj

δ

j

w

jdcij

+

Tu

C

c=1 N

i=1

uciloguci

+

Tv

C

c=1 D

j=1

v

cjlog

v

cj

+

Tw

D

j=1

w

jlog

δ

j

w

j

(12)

4

(5)

N.V. Pham, L.T. Pham, W. Pedrycz et al. Knowledge-Based Systems 216 (2021) 106549

To get optimal clustering results, the objective function FRFCoC (1)is minimized subject by the object membership functionsuci and the feature membership functions

v

cj are specific to each cluster and the feature weights

w

jto the following constraints:

C

c=1

uci

=

1

,

uci

[0

,

1]

, ∀

i

=

1

,

N

, ∀

c

=

1

,

C

D

j=1

v

cj

=

1

, v

cj

[0

,

1]

, ∀

c

=

1

,

C

, ∀

j

=

1

,

D

D

j

wj

=

1

,

wj

∈ [

0

,

1

] , ∀

j

=

1

,

D

(13)

That is,uciand

v

cjas analyzed in Eq.(8), Section2.1. And

w

jis the weight of each feature in the range [0,1] (

w

j

[0

,

1]

, ∀

j

=

1

,

D), the total weight of all features is 1.0 (

D

j=1

w

j

=

1).

Where, Tu,Tv and Tw are weights indicating fuzziness of U, V andW.

δ

jis used to adjust the feature weigh

w

j, the learning process for

δ

jis presented in Section2.1.

To minimize the objective function JFRFCoC with constraints are given by (1), we construct an objective function with C Lagrange coefficients

λ

c

(

c

=

1

,

C

)

corresponding toCclusters for constraint

C

c=1uci

=

1, andDLagrange coefficients

γ

j

(

j

=

1

,

D

)

corresponding toDfeatures for constraint

D

j=1

v

cj

=

1, andD Lagrange coefficients

ξ

j

(

j

=

1

,

D

)

corresponding toDfeatures for constraint

D

j=1

w

j

=

1, we obtain:

JFRFCoC(U

,

V

,

P

,

W

, δ

)

=

C

c=1 N

i=1 D

j=1

uci

v

cj

δ

j

w

jdcij

+

Tu

C

c=1 N

i=1

uciloguci

+

Tv

C

c=1 D

j=1

v

cjlog

v

cj

+

Tw

D

j=1

w

jlog

δ

j

w

j

+

C

c=1

λ

c

(

uci

1

)

+

D

j=1

γ

j

( v

cj

1

) +

D

j=1

ξ

j

( w

j

1

)

(14)

FRFCoC algorithm is resolved by the following steps. Firstly, we calculate the membership function Uby fixingV,P andW, then minimizing the objective function(14)according toU, and taking derivatives of objective function with respect to the fuzzy object memberships and setting them to zero to obtain,

JFRFCoC

uci

=

K

j=1

v

cj

δ

j

w

jdcij

+

Tu(loguci

+

1)

+ λ

c

=

0 (15) By doing some algebraic simplifications in(15), we obtain,

uci

=

e

D j=1

vcjδjwj dcij Tu

eλTuc

(16) Because of the constraint of

C

c=1uci

=

1, the Lagrange multiplier

λ

c is eliminated as follows

C

c=1

uci

=

C

c=1

e

D j=1

vcjδjwj dcij Tu

eλTuc

=

C c=1e

D j=1

vcjδjwj dcij Tu

eλTuc

=

1 (17)

eλTuc

=

C

c=1

e

D j=1

vcjδjwj dcij

Tu (18)

By using Eq.(18) in Eq. (16), the closed-form solution for the optimal object membership function is obtained as

uci

=

e

D j=1

vcjδjwj dcij Tu

C c=1e

D j=1

vcjδjwj dcij Tu

(19) Similarly, to find the optimal fuzzy feature membershipsV, taking derivatives of objective function with respect to the fuzzy feature memberships and setting them to zero, we obtain,

JFRFCoC

∂v

cj

=

C

c=1

uci

δ

jwjdcij

+

Tv(log

v

cj

+

1)

+ γ

j

=

0 (20) By some algebraic simplifications in Eq.(20), we reach,

v

cj

=

e

(N

i=1uciδjwj dcij

Tv )

D j=1e(

N i=1

uciδjwj dcij

Tv )

(21)

Before finding the cluster centroids P, the square of Euclidean distance

xij

pcj

2 is defined as

xij

pcj

2

=

(xij

pcj)2

=

x2ij

2xijpcj

+

p2cj, we obtain,

JFRFCoC

pcj

= v

cj

δ

j

w

j N

i=1

ucixij

− v

cj

δ

j

w

jpcj

N

i=1

uci

=

0 (22) Simplifying, we reach,

pcj

=

N i=1ucixij

N

i=1uci (23)

Next,

w

j is calculated by fixing U, V and P and the distances

xij

pcj

2are constants, then taking derivatives of the objective function (14)with respect to

w

j and setting them to zero, we obtain,

JFRFCoC

∂w

j

=

C

c=1 N

i=1

uci

v

cj

δ

jdcij

+

Tw(log

w

j

+

1)

+ ζ

j

=

0 (24) In the similar way toUandV, we easily obtain,

w

j

=

δ1je

C c=1

N

i=1ucivcjδj dcij Tw

D k=1

δ1ke

C c=1

N

i=1ucivckδk dcik Tw

(25)

FRFCoC algorithm diagram consists of the learning processes of membership function matrices U, V and feature weight matrix W that are shown as Algorithm3.

The output of FRFCoC algorithm consists of centroids and the distribution of data points in each cluster. For image data, clustering results are also presented as images by coloring pixels in each cluster with different colors. In addition, to compare and evaluate the performance of different algorithms, we also quantify the feature reduction results (the number of reduced features per each data set), the cluster quality evaluation indexes, the loops, and clustering time.

We next analyze the computational complexities for FRFCoC algorithms. This algorithm includes

τ

maxiterators to update com- ponents of the objective function(12): (1) Calculate membership functions,uci, which needs O(NC2D); (2) Calculate membership functions,

v

cj, which needsO(NCD2); (3) Update cluster centroids, pcj, which needsO(NC); (4) Update the weight

w

j, which needs O(NCD2). The total computational complexity isO(NC2D

+

NCD2), where N is the number of data patterns, C is the number of clusters, andDis the number of data dimensions. Normally, in the complex and multi-dimensions problems, D is larger than theC, therefore, the time complexity of FRFCoC is approximately

(6)

N.V. Pham, L.T. Pham, W. Pedrycz et al. Knowledge-Based Systems 216 (2021) 106549

Algorithm 3Feature-Reduction Fuzzy Co-Clustering algorithm Input: Data set X

= {

xi

,

xi

RD

}

,

i

=

1

..

N, the number of clusters C.

Output:The results of clustering and feature reduction: Centroids and the distribution of data points in each cluster and the number of reduction features.

1. Initialize parametersTu,Tv,Tw,

ε

1,

ε

2 and maximum number of iterations

τ

max.

2.Initializeuci,

w

jsuch that the constraint conditions in(13).

3.Set iteration

τ

=1.

4. REPEAT

5. Calculate

δ

jusing formula(5).

6. Updatepcjusing formula(23).

7. Calculatedcijusing Eq.(6).

8. Update

v

cjusing formula(21).

9. Updateuciusing formula(19).

10. Update

w

jusing formula(25).

11. For each

w

j(

τ

) inW(

τ

) If

w

j(

τ

)

< √

N

DThen 12. W(

τ

)=W(

τ

)

\ w

j(

τ

);

δ

(

τ

)=

δ

(

τ

)

\ δ

j(

τ

);

13. Remove the jth feature components fromX(

τ

), V(

τ

) andP(

τ

).

14. End For 15.

τ

=

τ

+1.

16. UNTIL

|∥

W

(τ) ∥ − ∥

W

(τ) ∥| < ε

or

τ ≤ τ

max

O(CD2N

τ

), i.e. similar to the FCoC algorithms. However, FRF- CoC algorithm reducesDafter each iteration, therefore, the time complexity of FRFCoC algorithm is smaller than FCoC algorithm.

Mathematically, FRFCoC algorithm has contributed a new ob- jective function and a learning scheme for clustering data. FRFCoC algorithm is a combination of advanced techniques of algorithms FRFCM and FCoC which simultaneously consider objects and fea- tures and reduce features in the machine learning schema. The objective function of FRFCoC not only considers the object mem- bership functions, but also considers the feature membership functions and the weight of features and the condition to remove the feature from the data processing loop. The schema of FRFCoC algorithm not only calculates the object membership functions and the feature membership functions, but also makes decisions to eliminate influential features to improve the performance in terms of accuracy and processing speed. To clarify the contribu- tions of FRFCoC algorithm, we would like to give some discussions below.

Compare with traditional clustering techniques, FRFCoC algo- rithm considers data in the detail such as the feature of data, the weight of the feature, and the ranking of the feature to eliminate the low-ranking features from the processing of data and improve the performance of clustering.

Compare with traditional fuzzy clustering algorithms, FRFCoC algorithm ranks the features of data and eliminated low-ranking features to reduce the features and improve the performance of clustering.

Compare with the previous reduction-feature techniques, which reduced the feature before conducting the data processing.

That is, the previous reduction-feature techniques are like a preprocessing step of data. Meanwhile, FRFCoC algorithm auto- matically calculates the parameters to reduce the feature during the data processing.

Compare with FRFCM algorithm, which is considered to be the same kind as FRFCoC algorithm. FRFCoC algorithm is formed on the basis of FCoC algorithm which the original algorithm

considered the feature and their weight. Therefore, integrating the reduction-feature technique into FCoC to form FRFCoC is very natural and achieving higher efficiency than FCoC algorithm is completely understandable and high logic meaning. Meanwhile, FRFCM algorithm itself is formed on the basis of FCM which only considers data by the object membership functions. There- fore, in order to form FRFCM, authors must integrate techniques of weighting of features and ranking features to eliminate the low-ranking features and improve the performance of clustering.

Logically, the integration of techniques in FRFCM shows its ratio- nality. However, FRFCM algorithm does not consider the feature membership function (as for FRFCoC) in ranking features that show the ranking features in FRFCM is weaker.

Based on the comments above, FRFCoC algorithm could achieve the performance better than previously proposed clus- tering algorithms including FRFCM algorithm which has recently proposed. It is essential to propose FRFCoC algorithm which demonstrates the science and novelty in automatically reducing the features and improving the performance of complex data clustering such as multi-dimensional, many-features data. To validate our predictions, this paper used theorems and lemmas to prove the convergence of FRFCoC algorithm (Section3.2) and conduct experiments (Section4) on the sample data sets and to demonstrate the performance of FRFCoC algorithm compared to previously proposed clustering algorithms.

3.2. Convergence theorems for FRFCoC algorithm

We next provide convergence theorems for the FRFCoC clus- tering algorithm that a FRFCoC convergent subsequence can tend to optimal solutions. Zangwill’s convergence theorem [29] and bordered Hessian matrix [36] will be used in our proof. We mention that this way had been used in Yang and Tian [37].

Originally, Zangwill defined a point-to-set map with T

:

V

P(V): where P(V) represents the power set of V and a closed point-to-set map must be defined. However, FRFCoC algorithm here is a point-to-point map and the ‘‘closed’’ property is exactly

‘‘continuity’’ for the case of the point-to-point map. Thus, the Zangwill’s convergence theorem is given as follows.

3.2.1. Zangwill’s convergence theorem

Let the point-to-point mapT

:

V

P(V) generate a sequence

{

zk

}

k=0 byzk+1

=

T

(

zk

)

. Let a solution setΩ

V be given and suppose that:

1. There is a continuous functionZ

:

V

R: such that, if zk

∈ /

,Z

(

T

(

zk

)) <

Z

(

zk

)

, and ifzk

∈ /

,Zk

(

T

(

zk

)) ≤

Z

(

zk

)

.

2. The mapT is continuous onV

\

.

3. All pointszkare contained in a compact setS

V.

Then the limit of any convergent subsequence shall be in the solution setΩ andZ(zk) will monotonically converge toZ(z) for somez

.

Set Mu

= {

U

=

[uci]CxN

C

c=1uci

=

1

,

uci

0

}

and Mv

= {

V

= [ v

cj

]

CxD

C

c=1

v

cj

=

1

, v

cj

0

}

andMw

= {

W

= [ w

j

]

D×1

D

j=1

w

j

=

1

, w

j

0

}

andP

= {

p1

,

p2

, . . . ,

pC

}

.

LetΩFRFCoC be a solution set for FRFCoC algorithm, defined as ΩFRFCOC

6

Referensi

Dokumen terkait

The 2-year Google-based Journal Impact Factor 2-GJIF is 2.01 2 International Journal of Economics and Management Systems This Journal receives many citations from the Web of Science