11
Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal c 2013 NSP Natural Sciences Publishing Cor. Semi-supervised Sparsity Pairwise Constraint Preserving Projections based on GA Mingming Qi 1,2 and Yang Xiang 1 1 Department of Computer Science and Technology, Tongji University, Shanghai 201804, P. R. China 2 School of Yuanpei, Shaoxing University, Shaoxing 312000, P. R. China Received: 15 Aug. 2012, Revised: 28 Nov. 2012, Accepted: 2 Jan. 2013 Published online: 1 May 2013 Abstract: The deficiency of the ability for preserving global geometric structure information of data is the main problem of existing semi-supervised dimensionality reduction with pairwise constraints. A dimensionality reduction algorithm called Semi-supervised Sparsity Pairwise Constraint Preserving Projections based on Genetic Algorithm (SSPCPPGA) is proposed. On the one hand, the algorithm fuses unsupervised sparse reconstruction feature information and supervised pairwise constraint feature information in the process of dimensionality reduction, preserving geometric structure in samples and constraint relation of samples simultaneously. On the other hand , the algorithm introduces the genetic algorithm to set automatically the weighted trade-off parameter for full fusion. Experiments operated on real world datasets show, in contrast to the existing typical semi-supervised dimensionality reduction algorithms with pairwise constraints and other semi-supervised dimensionality reduction algorithms on sparse representation, the proposed algorithm is more efficient. Keywords: Semi-supervised Dimensionality Reduction, Pairwise Constraints, Sparsity Preserving Projections, Information Fusion, Genetic Algorithm, Trade-off Parameter 1. Introduction The past few years have witnessed more and more research results on dimensionality reduction algorithms with pairwise constraints owing to great convenient in obtaining them and more supervised discriminant information in them [1,2]. Bar-Hillel et al. [3] proposed Constraints Fish Discriminant Analysis (CFDA), which is the pre-treatment step before relevant component analysis. CFDA only deal with the must-link constraints and ignore cannot-link constraints. Tang et al. [4] proposed Pairwise Constraints-guided Feature Projection (PCFP) for dimensionality reduction, which exploits both must-link constraints and cannot-link constraints but ignores unlabeled data. Zhang et al. [5] proposed Semi-Supervised Dimensionality Reduction (SSDR), which exploits both cannot-link and must-link constraints together with variance information in unlabeled data. However SSDR only deals with linear data. On the basic of LPP, Cevikalp et al. [6] proposed Constrained Locality Preserving Projections (CLPP), which finds neighborhood points to create a weighted neighborhood graph and make use of the constraints to modify the neighborhood relations and weight matrix to reflect this weak form of supervision. Although CLPP is fit for nonlinear data, the algorithm is still sensitive to noise and parameters. YU et al. [7] proposed Robust LPP (RLPP) with pairwise constraints based on robust path based similarity for overcoming these problems. Wei et al. [8] proposed Neighborhood Preserving based Semi-supervised Dimensionality Reduction (NPSSDR). The algorithm not only preserves the must-link and cannot-link constraints but also preserves the local structure of input data in the low dimensional embedding subspace by the regularization way, which makes NPSSDR easy to get into collapse in local structure. Chen et al. [9] proposed Semi-supervised Non-negative Matrix Factorization (SS-NMF) based on pairwise constraints on a few of documents. Peng et al. [10] proposed Semi-supervised Canonical Correlation Analysis Algorithm (Semi-CCA) which uses supervision information in the form of pairwise constraints in canonical correlation analysis. Davidson et al. [11] proposed Graph-driven Constrained Dimension Corresponding author e-mail: [email protected] c 2013 NSP Natural Sciences Publishing Cor.

Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065

Applied Mathematics & Information SciencesAn International Journal

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Semi-supervised Sparsity Pairwise ConstraintPreserving Projections based on GA

Mingming Qi1,2 and Yang Xiang1

1Department of Computer Science and Technology, Tongji University, Shanghai 201804, P. R. China2School of Yuanpei, Shaoxing University, Shaoxing 312000, P. R. China

Received: 15 Aug. 2012, Revised: 28 Nov. 2012, Accepted: 2 Jan. 2013Published online: 1 May 2013

Abstract: The deficiency of the ability for preserving global geometric structure information of data is the main problem of existingsemi-supervised dimensionality reduction with pairwise constraints. A dimensionality reduction algorithm called Semi-supervisedSparsity Pairwise Constraint Preserving Projections based on Genetic Algorithm (SSPCPPGA) is proposed. On the one hand, thealgorithm fuses unsupervised sparse reconstruction feature information and supervised pairwise constraint feature information in theprocess of dimensionality reduction, preserving geometric structure in samples and constraint relation of samples simultaneously.On the other hand , the algorithm introduces the genetic algorithm to set automatically the weighted trade-off parameter for fullfusion. Experiments operated on real world datasets show, in contrast to the existing typical semi-supervised dimensionality reductionalgorithms with pairwise constraints and other semi-supervised dimensionality reduction algorithms on sparse representation, theproposed algorithm is more efficient.

Keywords: Semi-supervised Dimensionality Reduction, Pairwise Constraints, Sparsity Preserving Projections, Information Fusion,Genetic Algorithm, Trade-off Parameter

1. Introduction

The past few years have witnessed more and moreresearch results on dimensionality reduction algorithmswith pairwise constraints owing to great convenient inobtaining them and more supervised discriminantinformation in them [1,2]. Bar-Hillel et al. [3] proposedConstraints Fish Discriminant Analysis (CFDA), which isthe pre-treatment step before relevant componentanalysis. CFDA only deal with the must-link constraintsand ignore cannot-link constraints. Tang et al. [4]proposed Pairwise Constraints-guided Feature Projection(PCFP) for dimensionality reduction, which exploits bothmust-link constraints and cannot-link constraints butignores unlabeled data. Zhang et al. [5] proposedSemi-Supervised Dimensionality Reduction (SSDR),which exploits both cannot-link and must-link constraintstogether with variance information in unlabeled data.However SSDR only deals with linear data. On the basicof LPP, Cevikalp et al. [6] proposed Constrained LocalityPreserving Projections (CLPP), which findsneighborhood points to create a weighted neighborhood

graph and make use of the constraints to modify theneighborhood relations and weight matrix to reflect thisweak form of supervision. Although CLPP is fit fornonlinear data, the algorithm is still sensitive to noise andparameters. YU et al. [7] proposed Robust LPP (RLPP)with pairwise constraints based on robust path basedsimilarity for overcoming these problems. Wei et al. [8]proposed Neighborhood Preserving basedSemi-supervised Dimensionality Reduction (NPSSDR).The algorithm not only preserves the must-link andcannot-link constraints but also preserves the localstructure of input data in the low dimensional embeddingsubspace by the regularization way, which makesNPSSDR easy to get into collapse in local structure. Chenet al. [9] proposed Semi-supervised Non-negative MatrixFactorization (SS-NMF) based on pairwise constraints ona few of documents. Peng et al. [10] proposedSemi-supervised Canonical Correlation AnalysisAlgorithm (Semi-CCA) which uses supervisioninformation in the form of pairwise constraints incanonical correlation analysis. Davidson et al. [11]proposed Graph-driven Constrained Dimension

∗ Corresponding author e-mail: [email protected]⃝ 2013 NSP

Natural Sciences Publishing Cor.

Page 2: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

1066 M. Qi, Y. Xiang: Semi-supervised Sparsity Pairwise Constraint...

Reduction via Linear Projection (GCDR-LP) that given aweighted graph attempts to find series of dimensions thatare linear combinations of the old dimensions.

Recently, sparse representations attain more and moreattention and are successfully applied in object detectionand classification [12, 13, 14, 15,16]. Researches show,classification based on sparse representations has goodrobustness on face datasets with deformities, dressingsand shelters. At present sparse learning has extended todimensionality reduction, including Sparse PrincipalComponent Analysis (SPCA) [17] Principal ComponentAnalysis with Weighted Sparsity Constraint (PCAWSC)[18] and Sparsity Preserving Projections (SPP) [19]. SPPis a new unsupervised dimensionality reductionalgorithm, which preserves sparse reconstruction relationsof high-dimension data to low-dimension data. In contrastto other unsupervised dimensionality reductionalgorithms, SPP preserves global geometry structureinformation contained in sparse reconstructions and isavailable of power discriminant analysis. However, SPP issensitive to variations in whole pattern of data owing todeficiencies of supervised information. Gu et al.[20]proposed Discriminative Sparsity PreservingProjections(DSPP). DSPP provides an explicit featuremapping by fitting the prior low-dimensionalrepresentations which are generated randomly by usingthe labels of the labeled data points and, meanwhile,setting the smoothness regularization term to measure theloss of the mapping in preserving the sparse structure ofdata, so DSPP has highly discriminative ability.

Motivated by above analyses, a Semi-supervisedSparse Pairwise Constraint Preserving Projections basedon Genetic Algorithm (SSPCPPGA) is proposed in thepaper. The algorithm firstly exact respectivelyunsupervised information of sparse reconstruction andsupervised information of pairwise constraint, then fusetwo kinds of information by the linear weighted way andseek the optimized weighted trade-off parameter throughthe genetic algorithm. Finally projections are gotten topreserve fused information. The projectedlow-dimensional data not only preserve global geometricstructure feature information contained in sparsereconstructions but also preserve pairwise constraintfeature information. Experimental results on AR, Yaleand UMIST show that our proposed algorithm improvesthe accuracy and stability of classification rate based onthe shortest Euclidean distance, in contrast to othertypical semi-supervised dimensionality reductionalgorithm with pairwise constraints and othersemi-supervised dimensionality reduction algorithms onsparse representation.

Several characteristics of our presented algorithm arelisted as follows:

(1) PCFP and SPP have their own advantages anddisadvantages. The fused algorithm inherits specialcharacter PCFP and SPP and overcomes theirdisadvantages.

(2)Sparse reconstruction information and pairwiseconstraint information vary greatly in different datasets,which is the reason that concrete weighted trade-offparameter is different in the process of linear fusion. Thealgorithm introduces the genetic algorithm to setautomatically the weighted trade-off parameter value inorder to get more performance.

The rest of the paper is organized as follows: Section2 reviews PCFP and SPP. SSPCPGA is introduced inSection 3. In Section 4, we compare proposed SSPCPGAwith PCFP, SPP, RLPP, SSDR, NPSSDR and DSPP. Theexperimental results are presented. Finally, we providesome concluding remarks and future work in Section 5.

2. Related works

In this section, we review related works, including PCFPin 2.1 and SPP in 2.2.

2.1. PCFP

Given training samples X = {x1,x2,x3, · · · ,xn} , must-linkset ML = {(xi,x j)|xi and x j are the same class} andcannot-link setCL = {(xi,x j)|xi and x j are not the same class} . PCFPaims to find an optimal projection matrix T thatmaximizes the following function[4]:

maxT

[∑

(xi,x j)∈CL∥ T T xi −T T x j ∥2 −

∑(xi,x j)∈ML

∥ T T xi −T T x j ∥2]

s.t. T T T = I

(1)

Where ∥ · ∥2 denotes the l2 norm.Eq.(1) may be understood in such a sentence that two

samples of must-link set in high-dimensional data spaceshould be more near in low-dimensional data space andtwo samples of cannot-link set in high-dimensional dataspace should be more further in low-dimensional dataspace.

2.2. SPP

Given training samples X = {x1,x2,x3, · · · ,xn} ∈ Rd×n ,sparse representation aims to reconstruct each sample xiwith else sample, using as few samples as possible. Asparse reconstructive weight vector si for each xi is gottenas follows:

minSi

∥ Si ∥1

s.t. xi = XSi

(2)

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 3: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) / www.naturalspublishing.com/Journals.asp 1067

(a) Projection on the initial dataset (b) Projection on the first changeddataset

(c) Projection on the second changeddataset

Figure 1 projected one-dimensional subspace

where Si j denotes the contribution of each x j toreconstructing xi . ∥ · ∥1 denotes the l1 norm .

On the basic of sparse representation, sparsereconstruction seeks a sparse reconstructive weight vectorxi for each xi through the following modified l1minimization problem:

minSi

∥ Si ∥1

s.t. xi = XSi

1 = 1T Si

(3)

where ∥Si∥1 denotes the l1 normal of Si ,Si = [Si1, · · · ,Sii−1,0,Sii+1, · · · ,Sin]

T ∈ Rd is a vector inwhich Si j denotes the contribution of each x j toreconstructing xi, and 1 ∈ Rn is a vector of all ones.

xi = Si1x1 + · · ·+Sii−1xi−1 +Sii+1xi+1

+ · · ·+Sinxn(4)

The sparse reconstruction matrix S = [S1,S2, · · · ,Sn]T

is attained through computing Si. Sparse reconstructivewei- ghts have intrinsic geometric properties of the data.

SPP aims to preserve sparse reconstruction relation ofhigh-dimensional data space to low-dimensional dataspace. Given the projection matrix T , T T XSi is theprojection point of xi in high-dimensional data space. Theobjective function of SPP is as follows[19]:

minT

n

∑i=1

∥T T xi −T T XSi∥2

s.t. T T XXT T = I

(5)

Eq.(5) can be further transformed to

maxT

[T T X(S+ST −ST S)XT T ]

s.t. T T XXT T = I(6)

3. Semi-supervised Sparse PairwiseConstraint Preserving Projections based onGA (SSPCPPGA)

In this section, we first introduce the basic idea of ouralgorithm and then the objective function is gotten; finallywe give steps of the algorithm.

3.1. Basic idea

In order to describe disadvantages of PCFP and SPP, wecreate a two-dimensional dataset that contains two classesrepresented by dot points and triangle points. The numberof two kinds of samples is same.Two kinds ofone-dimensional subspaces on the dataset and thechanged dataset are gotten through SPP and PCFP.Concrete results are shown in Figure 1.

In Figure 1, solid dot points denote labeled data of thefirst class and solid triangle points denote labeled data ofthe second class. The number of two kinds of labeledsamples is same. pairwise constraint sets are composed bythese labeled samples. Figure 1(a) shows projectionresults of SPP and PCFP. When labeled samples happento change in Figure 1(b),the projection result of SPP doesnot change while that of PCFP becomes poor. In contrastto data in Figure 1(a), data in Figure 1(c) are enlarged totwice as much in the vertical scalar to change globaldistribution patterns of the structure of data, which weakthe projection result of SPP and don’t affect that of PCFP.Figure 1 demonstrates that the performance of supervisedPCFP is sensitive to pairwise constraint sets instead ofglobal distribution patterns of the structure of trainingsamples and SPP is contrary. Therefore fusing featureinformation of PCFP and SPP in the process ofdimensionality reduction is a feasible way for overcomingdisadvantages of them.

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 4: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

1068 M. Qi, Y. Xiang: Semi-supervised Sparsity Pairwise Constraint...

Figure 2 The framework of setting the linear weighted parameter β based on GA

3.2. Objective function

According to the information level, information fusionsare divided into the data level, the feature level and thedecision-making level. Information fusion based on thefeature level may make sure relations among differentfeature information [21]. Fusion of supervised featureinformation and unsupervised feature information basedon the linear weighted way has proved to be an efficientfusion way of semi-supervised dimensionality reduction[22, 23].

SPCPPGA aims to find a projection T to preservesimultaneously pairwise constraint feature informationand sparse reconstruction feature information by makinguse of the linear weighted way to fuse two kinds ofinformation.

Eq.(1) and Eq.(6) can respectively further betransformed to an equivalent maximization problem asfollows:

maxT

[ ∑(xi,x j)∈CL

∥(T T xi −T T x j)∥2

T T T

−∑

(xi,x j)∈ML∥ (T T xi −T T x j) ∥2

T T T

] (7)

maxT

[T T X(S+ST −ST S)XT T ]T T XXT T

(8)

We infuse Eq.(7) and Eq.(8) to get the objectivefunction of SPCPPGA through the linear weighted way as

follows:

maxT

T T [βXSα XT +(1−β )Pα ]TT T [βXXT +(1−β )I]T

(9)

where

Sα = S+ST −ST S (10)

Pα = ∑(xi,x j)∈CL

(xi − x j)(xi − x j)T

− ∑(xi,x j)∈ML

(xi − x j)(xi − x j)T (11)

The weighted trade-off parameter β denotes thecontribution of sparse reconstruction feature informationto SPCPPGA and 1 − β denotes the contribution ofpairwise constraint feature information to SPCPPGA.

3.3. Optimize the weighted trade-off parameterbased on Genetic Algorithm

According to above analyses, the weighted trade-offparameter β plays an important role in SPCPPGA. Thenecessities of the optimization for the weighted trade-offparameter β are embodied as follows:

(1) Sparse reconstruction feature information is themodel of information decomposition and informationreconstruction that are based on over-complete dictionarywhile pairwise constraint feature information is side

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 5: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) / www.naturalspublishing.com/Journals.asp 1069

information. Their natures are great different owing totheir basic math ideas and representations.

(2) Two kinds of feature information from differenttraining samples and different dimensions of projectedsubspaces are very different , determining differentweighted trade-off parameters in the process of fusingthem.

Genetic Algorithm (GA) is the optimal solution bysimulating the natural evolutionary process search [24].GA has power of inherent implicated parallel and betterability of global optimization, obtaining automaticallyand guiding to optimize the searching space throughadjusting adaptively searching directions [25, 26].Therefore, GA is introduced to guide to set the linearweighted trade-off parameter . Figure2 gives us theframework of setting the linear weighted trade-offparameter based on GA.

The main processes of optimizing the weighted trade-off parameter β in Figure 2 are as follows:

(1) Create randomly a certain number of populationsrepresented by the binary chromosome and initialize theirvalues between 0 and 1.

(2) Achieve basic operation on these populations,including selecting, crossing and mutating. Selectingoperation is achieved by eliminating the first half ofindividuals in ascending order about the fitness value andcreating randomly some individuals for supplement.

(3) Get most optimized value in the generation and theglobal optimized value in the total generation.

(4) If the number of GA generation g < 4, then jumpto (1), otherwise output the global optimized value in allthe generation.

3.4. Algorithm steps

Input: training samples X = {x1,x2,x3, · · · ,xn} andxi ∈ Rw×h ,ML = {(xi,x j)|xi and x j are in the same class} ,CL = {(xi,x j)|xi and x j are not in the same class}, Thesize of the populations p , the length of the chromosomec, the crossover probability pc and mutation probabilitypm.Output:the project matrix T d×r(r < d).Steps:

(1) construct S using to Eq.(3).(2) construct Sα using Eq.(10).(3) construct Pα using Eq.(11).(4)The simplest Nearest Neighbor Classifier is adopted

in the fitness function. According to the model of Figure2,calculate the optimized weighted parameters with differentdatasets and different dimensions.

(5)Transform Eq.(9) to

(βXSα XT +(1−β )Pα)φ = λ (βXXT +(1−β )I)φ

and calculate the projections matrix T d×r(r < d).

4. Experiments and analyses

In this section, we firstly introduce experimental datasets,which is followed by experimental settings. Finally, wegive experimental results and detail analyses on theclassification performance of SPCPPGA and theadaptivity of the weighted trade-off parameter β .

4.1. Experimental datasets

(1) Yale contains 165 face images of 15 individuals.There are 11 images per subject, and these 11 images arerespectively, under the following different facialexpression or configuration: center-light, wearing glasses,happy, left-light, wearing no glasses, normal, right-light,sad, sleepy, surprised and wink. In our experiment, weresize theses face images of Yale to 30×30 pixels.

(2) AR consists of over 4000 face images of 126individuals. For each individual, 26 pictures were taken intwo sessions that separated by two weeks and eachsection contains 13 images, which include front view offaces with different expressions, illuminations andocclusions. In our experiment, we resize theses faceimages of AR to 30×30 pixels.

(3) UMIST is composed of 564 face images of 20individuals. UMIST face images cover the front face ofdifferent side. In our experiment, we resize theses faceimages of UMIST to 34×28 pixels.

A group of face samples on Yale, AR and UMIST areshown in Figure3.

(a) A group of face samples on Yale

(b) A group of face samples on AR

(c) A group of face samples on UMIST

Figure 3 A group of face samples on various datasets

4.2. Experimental settings

SSDR[5], RLPP[7],NPSSDR[8] and DSPP [20] are threetypical semi-supervised dimensionality reductionalgorithms with pairwise constraints and are comparedwith our proposed algorithm for testing classification

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 6: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

1070 M. Qi, Y. Xiang: Semi-supervised Sparsity Pairwise Constraint...

performances. In addition, SPP and PCFP are also addedcompared algorithms in order to verify the integrationperformance. Table 1 shows specific parameter settings invarious algorithms.

Table 1 Specific settings of parameters on algorithmsAlgorithm name Parameters settings

SPP noPCFP no

SSPCPPGA(p,c) pc = 0.6, pm = 0.1RLPP κ = 7, t = 1SSDR α = 1,β = 20

NPSSDR α = 0.1,κ = 5DSPP no

In Table 1,p denotes the population size, c denotes theindividual chromosome length, pc denotes the crossoverprobability and pm denotes the mutation probability inSSP CPPGA. Usually the crossover probability parameterpc and the mutation probability parameter pm of GA areset respectively to 0.6 and 0.1. For expressionconvenience, SSPCPPGA with the different populationsize and the different individual chromosome length isdenoted by SSPC PPGA (p,c).

The simplest Nearest Neighbor classificationalgorithm is adopted. We select randomly L images fromeach group face for training samples and remains fortesting. Retained feature dimensions are increased withthe increment D and corresponding classificationaccuracies are calculated. All experiments are repeated 20times and average recognition rates are gotten.

Besides,the matrix XXT probably is singular since thenumber of training samples is much smaller than thefeature dimension. To deal with the problem, training datafirstly projected into the PCA subspace X̃ = T T

PCAX . Theratio of PCA is set to 1.

4.3. Experimental results and analyses

4.3.1. Results and analyses on the classificationperformance of SSPCPPGA

In the experiment, we select SSPCPPGA(5,10) where theparameter p is set to 5 and the parameter c is set to 10.Fig.4-Fig.6 show experimental results of the classificationperformance with L training samples of each group and Dincrement of retained dimensions.

Moreover, in order to verify more accurately theclassification performance of SSPCPPGA, the mostrecognition rates of various algorithms on different facedatasets under different L are given and shown in Table2.Bold items represent maximum recognition rates.

From above Fig.7-Fig.9 and Table 2, we drawfollowing conclusions:

(1) SSPCPPGA is obviously superior to SPP. Besides,SSPCPPGA outperforms PCFP and also can get top

recognition rate under lower dimensions. This illuminatesthat the linear weighted fusion way in Eq(10) is efficientand SSPCPPGA inherits advantages of SPP and PCFP.

(2) In contrast to SSDR, the advantages ofSSPCPPGA is obvious. The reason is that SSDRpreserves structure information based on linear scattermatrices while SSPCPPGA adopts to preserve structureinformation with sparse representation that has morepower for describing geometric structures in data.

(3) Although RLPP and SSPCPPGA share nonlinearadvantages, SSPCPPGA outperforms RLPP owing muchto different description ways for manifold structure indata. RLPP only capture local manifold structures withthe nonlinear approximation, which ignore naturalintrinsic geometric structure information in data whileSparse reconstruction of SSPCPPGA is available ofintrinsic geometric properties.

(4) NPSSDR preserves local structure information byadding a regularization item instead of constructing theadjacency matrix of data. However, NPSSDR does notchange the way of preserving structure information withthe nonlinear approximation. Therefore the performanceof NPSSDR is not so good as that of SSPCPPGA.

(5) DSPP has a high discriminative ability which isinherited from the sparse representation of data that isshared by DSPP and SSPCPPGA. But DSPP attempts tomaintain the prior low-dimensional representationconstructed by the data points and the known class labels,which is the main reason for that SSPCPPGA isobviously superior to DSPP.

4.3.2. Analyses on the effect of parameters of GA on thealgorithm

To evaluate the effect of the parameter p and the parameterc on the algorithm, we set different p and c in experiments.Experimental results on the effect of parameters are shownin Fig.7-Fig.9.

From experimental results of Fig.7-Fig.9, we draw theconclusion that the performance of SSPCPPGA is notsensitive to the parameter p and the parameter c whenthey exceed the threshold value.

4.3.3. Analyses on the optimization performance of theweighted trade-off parameter β

In order to evaluate the adaptivity of the weightedtrade-off parameter β , we firstly select some retaindimension and increase gradually β from 0 to 1 with theincrement of 0.05 and calculate correspondingrecognition rate. Moreover, the optimized weightedtrade-off parameter β of different retained dimension arelisted, which are obtained by SSPCPPGA(5,10).Fig.10-Fig.12 show experimental results on AR,Yale andUMIST.

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 7: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) / www.naturalspublishing.com/Journals.asp 1071

(a) L=5 and D=10 (b) L=10 and D=10

Figure 4 Experimental results of the classification performance with L and D on AR

(a) L=3 and D=2 (b) L=6 and D=5

Figure 5 Experimental results of the classification performance with L and D on Yale

(a) L=3 and D=2 (b) L=6 and D=5

Figure 6 Experimental results of the classification performance with L and D on UMIST

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 8: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

1072 M. Qi, Y. Xiang: Semi-supervised Sparsity Pairwise Constraint...

Table 2 The most recognition rates of various algorithms on different face datasets under different L

Algorithm nameFace datasets

AR Yale UMIST5 10 3 6 3 6

SPP 72.55% 85.53% 63.33% 81.00% 57.03% 79.81%PCFP 71.71% 80.04% 78.75% 78.33% 80.79% 88.08%

SSPCPPGA(5,10) 82.15% 88.39% 89.38% 91.04% 87.34% 91.83%RLPP 71.73% 80.02% 78.75% 78.33% 81.20% 85.38%SSDR 70.71% 77.97% 77.71% 78.33% 79.77% 81.62%

NPSSDR 79.73% 87.34% 81.96% 85.67% 81.77% 87.92%DSPP 77.01% 85.53% 82.96% 86.67% 83.94% 86.90%

(a) L=5 and D=10 (b) L=10 and D=10

Figure 7 Experimental results of the effect of parameters with L and D on AR

(a) L=3 and D=2 (b) L=6 and D=5

Figure 8 Experimental results of the effect of parameters with L and D on Yale

From experimental results of Fig.10-Fig.12, we drawfollowing conclusions:

(1) When retained dimension number exceeds thethreshold of dimension, the role of the weighted trade-offparameter β on the classification performance is nearlyidentical, which illuminates the stability of the adaptivityof the weighted trade-off parameter β inSSPCPPGA(5,10).

(2) The way of interval of 0.05 increments need tocalculate 20 individual value to obtain the optimizedvalue while SSPCPPGA(5,10) only to calculate 15individual value, which demonstrates less time cost ofSPCPPGA in searching the optimized value.

(3) The optimized trade-off parameter β calculated bySSPCPPGA(5,10) is almost near to the optimal value inexperimental results, which confirm its effectiveness of the

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 9: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) / www.naturalspublishing.com/Journals.asp 1073

(a) L=3 and D=2 (b) L=6 and D=5

Figure 9 Experimental results of the effect of parameters with L and D on UMIST

(a) L=5 (b) L=10

Figure 10 Experimental results of the optimization of the weighted trade-off parameter β with L on AR

(a) L=3 (b) L=6

Figure 11 Experimental results of the optimization of the weighted trade-off parameter β with L on Yale

adaptivity of the weighted trade-off parameter β inSSPCPPGA(5,10).

(4) The optimized trade-off parameter β is differenton different datasets, which demonstrates the necessity of

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 10: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

1074 M. Qi, Y. Xiang: Semi-supervised Sparsity Pairwise Constraint...

(a) L=3 (b) L=6

Figure 12 Experimental results of the optimization of the weighted trade-off parameter β with L on UMIST

the adaptivity of the weighted trade-off parameter β inSSPCPPGA(5,10).

4.4. Computational complexity analyses

For samples X = {x1,x2,x3, · · · ,xn} ∈ Rd×n ,SSPCPPGAcontains main steps for solving Sa,Pa ,the trade-off βbased on GA and the eigen-decomposition using Eq.(10).The computational complexity of sparse learning is nearlythat of solution of l1 norm minimization problems whichis O(d3) [27]. Therefore the computational complexity ofsolving Sa is O(d3). The computational complexity ofsolving Pa is O(d2).The eigen-problem on a symmetricmatrix can be efficiently computed by the singular valuedecomposition (SVD) which is O(d3). Maincomputational time in GA algorithm is caused bycalculating the adaptability that contains steps for solvingSa, Pa, and the eigen-decomposition. So thecomputational complexity of solving the trade-off βbased on GA is O(d3). To sum up, the computationalcomplexity of SSPCPPGA is O(d3). Computationalcomplexity analyses on various algorithms are shown inTable 3.

Table 3 Computational complexity analysesAlgorithm name Computational complexity

SPP O(d3)PCFP O(d3)

SSPCPPGA O(d3)DSPP O(d3)RLPP O(d3)SSDR O(d3)

NPSSDR O(d3 +n2)

Conclusion

In the paper, a dimensionality reduction algorithm calledSemi-supervised Sparsity Pairwise Constraint PreservingProjections based on Genetic Algorithm (SSPCPPGA) isproposed to solve the problem of deficiency of the abilityfor preserving global geometric structure of data inexisting semi-supervised dimensionality reduction withpairwise constraints. On one hand, the algorithm fusessparse reconstruction feature information and pairwiseconstraint feature information through the linear weightedway. Projected data preserve geometrical structure insamples and constraints relation of samples. On the otherhand, the algorithm adopts GA to get the optimizedweighted trade-off parameter. Experiments operated onAR, Yale and UMIST demonstrate the effectiveness ofour proposed algorithm.

However, to optimize weighted trade-off parameterbased on GA still costs us some time, the fusion waywithout selecting parameters is more convenience, whichis the future work.

Acknowledgments

The research is supported by NSF of China(grant No.71171148) and NSF of Zhejiang province(grant No.LQ12F02007, Y201122544).

References

[1] R. Yan, J. Zhang, J. Yang and A. Hauptmann, Adiscriminative learning framework with pairwise constraintsfor video object classification, Proceedings of the IEEEcomputer society conference on computer vision and patternrecognition, 2,284-291(2004).

[2] E. Xing, A. Ng, M. Jordan and S. Russell, Distance metriclearning with application to clustering with side information,

c⃝ 2013 NSPNatural Sciences Publishing Cor.

Page 11: Semi-supervised Sparsity Pairwise Constraint Preserving ... · Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) 1065 Applied Mathematics & Information Sciences An International Journal

Appl. Math. Inf. Sci. 7, No. 3, 1065-1075 (2013) / www.naturalspublishing.com/Journals.asp 1075

Advances in Neural Information Processing Systems, MITPress, Cambridge, MA, 15, 505-512 (2003).

[3] A. Bar-Hillel, T. Hertz, N. Shental and D. Weinshall,Learning a mahalanobis metric from equivalence constraints,Journal of Machine Learning Research, 6(6), 937-965 (2006).

[4] W. Tang and S. Zhong, Pairwise constraints-guideddimensionality reduction, SDM Workshop on FeatureSelection for Data Mining, 2006.

[5] D. Zhang, Z. Zhou and S. Chen, Semi-Superviseddimensionality reduction, Proceedings of the SDM, 629-634(2007).

[6] H. Cevikalp, J. Verbeek, F. Jurie and A. Klaser, Semi-Supervised dimensionality reduction using pairwiseequivalence constraints, Proceedings of the VISAPP,489-496 (2008).

[7] G. Yu, H. Peng, J. Wei and Q. Ma, Robust LocalityPreserving Projections with Pairwise Constraints, Journalof Computational Information Systems, 6(5), 1631-1636(2010).

[8] J. Wei and H. Peng,Neighbourhood preserving based semi-supervised dimensionality reduction,Electronics Letters,44(20), 1190-1191 (2008).

[9] Y. Chen, M. Rege, M. Dong and J. Hua, Incorporating userprovided constraints into document clustering, Proceedingsof the ICDM, 103-112 (2007).

[10] Y. Peng and D. Q. Zhang, Semi-Supervised canonicalcorrelation analysis algorithm, Journal of Software, 19(11),2822-2832 (2008).

[11] I. Davidson, Knowledge driven dimension reduction forclustering, Proceedings of the IJCAI, 1034-1039 (2009).

[12] M. Davenport, M. Duarte, M. Wakin, D. Takhar, K. Kellyand R. Baraniuk, The smashed filter for compressiveclassification and target recognition, Proceedingsof IS&T/SPIE Symposiumon Electronic Imaging:Computational Imaging, 2007.

[13] J. Wright, A. Yang, S. Sastry and Y. Ma, Robust facerecognition via sparse representation, IEEE Transaction onPattern Analysis and Machine Intelligence, 31(2), 210-227(2009).

[14] T. Guha, and R.K. Ward, Learning Sparse Representationsfor Human Action Recognition, IEEE Transactions on PatternAnalysis and Machine Intelligence, 34(8), 1576-1588 (2012).

[15] Y. Liu and W. Wei, Sparse Kernel Learning-based NonlinearPredictive Controllers, Applied Mathematics & InformationSciences, 5(2), 195-201 (2011).

[16] S. Wang, L.J. Cui ,D.C. Liu, R. Huck, P. Verma, J.J. Sluss, Jr. And S. Cheng, Vehicle Identification ViaSparse Representation, IEEE Transactions on IntelligentTransportation Systems, 13(2), 955-962 (2012).

[17] H. Zou, T. Hastie and R. Tibshirani, Sparse principalcomponent analysis,Technical report, 2004.

[18] T. D. X. Duong, and V. N. Duong, Principal ComponentAnalysis with Weighted Sparsity Constraint, AppliedMathematics & Information Sciences, 4(1), 79-91 (2010).

[19] L. Qiao, S. Chen and X. Tan, Sparsity preserving projectionswith applications to face recognition, Pattern Recognition,43(1),331-341(2010).

[20] N. N. Gu, M. Y. Fan, H. Qiao and BoZhang, DiscriminativeSparsity Preserving Projections for Semi-SupervisedDimensionality Reduction, IEEE Signal Processing Letters,19(7), 391-394 (2012).

[21] J. Esteban, A. Starr and R. Willets, A Review of Data Fusionmodels and Architectures: towards engineering guidelines,Neural Computing & Applications, 14(4), 273-287 (2005).

[22] Y. Q. Song, F. P. Nie , C. S. Zhang and S. Xiang ,A unified framework for semi-supervised dimensionalityreduction,Pattern Recognition, 41(9), 2789-2799 (2008).

[23] M. Sugiyama, Dimensionality reduction of multimodallabeled data by local Fisher discriminant analysis, Journal ofMachine Learning Research, 8(5), 1027-1061 (2007).

[24] K. J. Leng and X.J. Chen, A Genetic AlgorithmApproach for TOC-based Supply Chain Coordination,Applied Mathematics & Information Sciences, 6(3), 767-774(2012).

[25] A. Isazadeh, H.Izadkhah and A.H. Mokarram, A Learningbased Evolutionary Approach for Minimization of MatrixBandwidth Problem, Applied Mathematics & InformationSciences, 6(1), 51-57 (2012)

[26] J. C. Hung, C. C. Lin, Hybridization of Genetic Algorithmand IBMUSIC applied to DOA estimation under coherentsignals, Applied Mathematics & Information Sciences, 6(2),489-496 (2012).

[27] D. L. Donoho and Y. Tsaig, Fast Solution of l1-norm Minimization Problems When the Solution May beSparse,IEEE Transactions on Information Theory,54(11),4789-4812 (2008).

MingMing Qi was bornin Jiangxi, China, in 1974.He received the MS degreein computer application fromGuizhou University in 2003.Currently he is a Ph.D.studentat the Department ofComputer Science andEngineering in Tongji

university and a lecturer in Shaoxing university. Hisresearch interests are in the areas of machine learning andimage processing.

Yang Xiang wasborn in Shandong, China. Hereceived Ph.D. from DalianUniversity of Technology.Currently he is a prefessor ofthe Department of ComputerScience and Engineeringin Tongji university. Hisresearch interests are in the

areas of machine learning ,web semantic and imageprocessing.

c⃝ 2013 NSPNatural Sciences Publishing Cor.