9
Research Article An Analysis Dictionary Learning Algorithm under a Noisy Data Model with Orthogonality Constraint Ye Zhang, 1 Tenglong Yu, 1 and Wenwu Wang 2 1 Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China 2 Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford GU2 7XH, UK Correspondence should be addressed to Ye Zhang; [email protected] Received 25 January 2014; Revised 13 June 2014; Accepted 26 June 2014; Published 13 July 2014 Academic Editor: Francesco Camastra Copyright © 2014 Ye Zhang et al. is is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Two common problems are oſten encountered in analysis dictionary learning (ADL) algorithms. e first one is that the original clean signals for learning the dictionary are assumed to be known, which otherwise need to be estimated from noisy measurements. is, however, renders a computationally slow optimization process and potentially unreliable estimation (if the noise level is high), as represented by the Analysis K-SVD (AK-SVD) algorithm. e other problem is the trivial solution to the dictionary, for example, the null dictionary matrix that may be given by a dictionary learning algorithm, as discussed in the learning overcomplete sparsifying transform (LOST) algorithm. Here we propose a novel optimization model and an iterative algorithm to learn the analysis dictionary, where we directly employ the observed data to compute the approximate analysis sparse representation of the original signals (leading to a fast optimization procedure) and enforce an orthogonality constraint on the optimization criterion to avoid the trivial solutions. Experiments demonstrate the competitive performance of the proposed algorithm as compared with three baselines, namely, the AK-SVD, LOST, and NAAOLA algorithms. 1. Introduction Sparse signal representation has been the focus of much recent research in signal processing fields such as image denoising, compression, and source separation. Sparse rep- resentation is oſten established on the synthesis model [13]. Considering a signal x , the synthesis sparse representation of x can be described as x = Da with a0 = , (1) where ‖⋅‖ 0 represents the 0 pseudonorm, defined as the number of nonzero elements of a vector. D × is a possibly overcomplete dictionary (), and a , containing the coding coefficients, is assumed to be sparse with . Recently, an alternative form of sparse representation model called analysis model was proposed in [414]. In this model, an overcomplete analysis dictionary or operator Ω × () is sought to transform x to a high dimensional space; that is, Ωx = z with z0 = − , (2) where z is called the analysis representation of x and assumed to be sparse and is cosparsity of the analysis model, which is the number of zeros in the vector z. Both models can be used to reconstruct an unknown signal x from the corrupted measurement y ; that is, y = x + k, (3) where k is a Gaussian noise vector. is is an ill-posed linear inverse problem, which has been studied extensively [15, 16]. Using the synthesis model, the original signal x can be estimated by solving a = arg min a a0 subject to Da 2 2 (4) and then calculated as x = D a, where denotes the noise floor. It is now well known [6] that the original signal can also Hindawi Publishing Corporation e Scientific World Journal Volume 2014, Article ID 852978, 8 pages http://dx.doi.org/10.1155/2014/852978

Research Article An Analysis Dictionary Learning Algorithm

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Research Article An Analysis Dictionary Learning Algorithm

Research ArticleAn Analysis Dictionary Learning Algorithm under a Noisy DataModel with Orthogonality Constraint

Ye Zhang1 Tenglong Yu1 and Wenwu Wang2

1 Department of Electronic Information Engineering Nanchang University Nanchang 330031 China2 Centre for Vision Speech and Signal Processing University of Surrey Guildford GU2 7XH UK

Correspondence should be addressed to Ye Zhang zhangyencueducn

Received 25 January 2014 Revised 13 June 2014 Accepted 26 June 2014 Published 13 July 2014

Academic Editor Francesco Camastra

Copyright copy 2014 Ye Zhang et alThis is an open access article distributed under the Creative CommonsAttribution License whichpermits unrestricted use distribution and reproduction in any medium provided the original work is properly cited

Two common problems are often encountered in analysis dictionary learning (ADL) algorithms The first one is that the originalclean signals for learning the dictionary are assumed to be known which otherwise need to be estimated from noisymeasurementsThis however renders a computationally slow optimization process and potentially unreliable estimation (if the noise level ishigh) as represented by the Analysis K-SVD (AK-SVD) algorithm The other problem is the trivial solution to the dictionary forexample the null dictionary matrix that may be given by a dictionary learning algorithm as discussed in the learning overcompletesparsifying transform (LOST) algorithm Here we propose a novel optimization model and an iterative algorithm to learn theanalysis dictionary where we directly employ the observed data to compute the approximate analysis sparse representation of theoriginal signals (leading to a fast optimization procedure) and enforce an orthogonality constraint on the optimization criterionto avoid the trivial solutions Experiments demonstrate the competitive performance of the proposed algorithm as compared withthree baselines namely the AK-SVD LOST and NAAOLA algorithms

1 Introduction

Sparse signal representation has been the focus of muchrecent research in signal processing fields such as imagedenoising compression and source separation Sparse rep-resentation is often established on the synthesis model [1ndash3] Considering a signal x isin 119877

119872 the synthesis sparserepresentation of x can be described as

x = Da with a0 = 119896 (1)

where sdot0represents the ℓ

0pseudonorm defined as the

number of nonzero elements of a vector D isin 119877119872times119873 is a

possibly overcomplete dictionary (119873 ge 119872) and a isin 119877119873containing the coding coefficients is assumed to be sparsewith 119896 ≪ 119873

Recently an alternative form of sparse representationmodel called analysis model was proposed in [4ndash14] In thismodel an overcomplete analysis dictionary or operator Ω isin

119877119875times119872 (119875 ge 119872) is sought to transform x isin 119877119872 to a high

dimensional space that is

Ωx = z with z0 = 119875 minus 119897 (2)

where z isin 119877119875 is called the analysis representation of x andassumed to be sparse and 119897 is cosparsity of the analysis modelwhich is the number of zeros in the vector z

Both models can be used to reconstruct an unknownsignal x from the corrupted measurement y isin 119877119872 that is

y = x + k (3)

where k isin 119877119872 is a Gaussian noise vector This is an ill-posedlinear inverse problem which has been studied extensively[15 16] Using the synthesis model the original signal x canbe estimated by solving

a = argmina

a0 subject to 1003817100381710038171003817119910 minusDa100381710038171003817

1003817

2

2le 120576 (4)

and then calculated as x = Da where 120576 denotes the noisefloor It is nowwell known [6] that the original signal can also

Hindawi Publishing Corporatione Scientific World JournalVolume 2014 Article ID 852978 8 pageshttpdxdoiorg1011552014852978

2 The Scientific World Journal

be recovered by solving the following analysis model basedoptimization problem

x = arg min Ωx0 subject to 1003817100381710038171003817y minus x100381710038171003817

1003817

2

2le 120576 (5)

For the squared and invertible dictionary the synthesis andthe analysis models are identical withD = Ωminus1 [9] Howeverthe two models depart if we concentrate on the redundantcase (119875 gt 119872 and 119873 gt 119872) In the analysis model the signalx is described by the zero elements of z that is zero entriesin z that define a subspace which the signal x belongs toas opposed to the few nonzero entries of a in the synthesismodel

The performance of both the synthesis and analysismodels hinges on the representation of the signals with anappropriately chosen dictionary In the past decade a greatdeal of effort has been dedicated to learning the dictionary forthe synthesis model however the ADL problem has receivedless attention with only a few algorithms proposed recently[5 6 10 17ndash19] In these works the dictionary Ω is oftenlearned from the observed signals Y = [y

1 y2 y

119870] isin

119877119872times119870 measured in the presence of additive noise that is

Y = X + V (6)

where X = [x1 x2 x

119870] isin 119877

119872times119870 contains the originalsignals V = [k

1 k2 k

119870] isin 119877

119872times119870 is a Gaussian noisematrix and 119870 is the number of signals

Exploiting the fact that a row of the dictionary Ω isorthogonal to a subset of training signals X a sequentialminimal eigenvalue based ADL algorithm is proposed in [5]Once the subset is found the corresponding row of the dic-tionary can be updated with the eigenvector associated withthe smallest eigenvalue of the autocorrelation matrix of thesesignals However as 119875 increases so does the computationalcost of the method In [6ndash8] an ℓ

1-norm penalty function

is applied to Ωx and a projected subgradient algorithm(called NAAOLA) is proposed for analysis operator learningThese works employ a uniformly normalized tight frame asa constraint on the dictionary to avoid the trivial solution(eg Ω = 0) which however limits the range of possibleΩ to be learned In [10] the Analysis K-SVD (AK-SVD)algorithm is proposed for ADL By keeping Ω fixed theoptimal backward greedy algorithm (OBG) is employed toestimate a submatrix of Ω whose rows are then used todetermine a submatrix of Y and the eigenvector associatedwith the smallest eigenvalue of this submatrix is then used toupdateΩ A generalized analysismodel that is the transformmodel is proposed in the learning sparsifying transform(LST) algorithm [11] Unlike the analysismodel (2) the sparserepresentation of a signal in the LSTmodel is not constrainedto lie in the range space ofΩ Such a generalization allows thetransform model to accommodate a wider class of signals Aclosed-form solution to the LST problem is also developedin [13] The LST algorithm [11] is further extended to theovercomplete case leading to the LOST algorithm in [12]The transform K-SVD algorithm proposed recently in [14] isessentially a combination of the ideas in the LOST and AK-SVD algorithms

There are two problems that have been observed oralready studied in some of the ADL algorithms discussedabove The first one is associated with the computationalcost in the optimization process of the ADL algorithmsFor example in the AK-SVD algorithm X needs to beestimated before learning the dictionary and as a result theoptimization process becomes computationally demandingThe second one is associated with the trivial solutions to thedictionary which may lead to spurious sparse representa-tions To eliminate such trivial solutions extra constraints arerequired such as the full-rank constraint on Ω employed in[11ndash13] and the mutual coherence constraint on the rows ofΩin [12 14]

In this paper we propose a new optimization model andalgorithm attempting to provide alternative solutions to thetwo potential problems mentioned above More specificallysimilar in spirit to our recent work [17] we directly use theobserved data to compute the approximate analysis sparserepresentation of the original signals without having topreestimate X for learning the dictionary This leads to acomputationally very efficient algorithm Moreover differentfrom the LOST algorithm we enforce an orthogonalityconstraint onΩ which as shown later has led to an improvedlearning performance

The paper is organized as follows In Section 2 we discussthe novel ADL model and algorithm In Section 3 we showsome experimental results before concluding the paper inSection 4

2 The Proposed ADL Algorithm

In some algorithms discussed above such as [10] the opti-mization criterion is based on X In practice however Xis unknown and required to be estimated from Y Thisunfortunately results in a computationally very slow opti-mization process as shown in Section 31 To address thisissue we introduce a new model and algorithm for ADLwhere similar to [17] we use ΩY

1as an approximation to

ΩX1 which is further replaced by Z

1and then used as

a new constraint for the reconstruction term Z = ΩY Thisleads to the following model of ADL

min(12

Z minusΩY2119865+ 120582Z1) (7)

where 120582 gt 0 is a regularization parameter According tothe analysis model (2) Z would be an analysis coefficientmatrix only if its columns lie in the range space of the analysisdictionary Ω However Z=ΩY is not explicitly enforced in(7) and hence it is called the transform coefficient matrix asin [19] Using (7) the analysis dictionaryΩ is learned from Ywithout having to preestimate X as opposed to the AK-SVDalgorithm so the computational effort for estimating X fromY is exempted

With the model (7) however there is a trivial solutionthat is Ω = 0 Z = 0 To avoid such a solution additionalconstraints on Ω are required In [12] a full column rankconstraint has been imposed on Ω through the use of anegative log determinant of Ω119879Ω However the inverse of

The Scientific World Journal 3

Table 1 Image denoising results (PSNR in dB)

120590 Noisy in dB ADL method Lena House Peppers

5 3415

OIHT-ADL 3826 3808 3770NAAOLA 3736 3704 3593AK-SVD 3844 3920 3793LOST 3816 3850 3747

10 2813

OIHT-ADL 3502 3479 3389NAAOLA 3273 3261 3112AK-SVD 3485 3532 3382LOST 3475 3477 3361

15 2461

OIHT-ADL 3317 3305 3169NAAOLA 3087 3082 2922AK-SVD 3259 3298 3128LOST 3288 3282 3138

20 2211

OIHT-ADL 3183 3171 3013NAAOLA 2965 2922 2740AK-SVD 3138 3151 2976LOST 3164 3147 2984

Ω119879Ω needs to be computed in the conjugate gradientmethod

in [12] and the inverse of Ω119879Ω may affect the stability ofthe numerical computation when the initial of Ω119879Ω is illconditioned To address this problem in our work a functiondefined on Ω119879Ω = I is employed as a constraint term whichenforcesΩ to be a full column rank matrix based on the factthat the ranks of Ω and its corresponding Gram matrix areequal This leads to the following new optimization criterion

min(12

Z minusΩY2119865+ 120582Z1) subject to Ω119879Ω = I (8)

Using a Lagrangian multiplier 120574 the optimization problemcan be reformulated as

min(12

Z minusΩY2119865+ 120582Z1 +

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865) (9)

It is worth noting that our proposed objective function(9) is essentially different from the one used in [7] In [7] theobjective function is defined as follows

min(ΩX1 +120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865+

120582

4

sum

119894

1003817100381710038171003817100381710038171003817

w119879119894w119894minus

119872

119875

1003817100381710038171003817100381710038171003817

2

2

) (10)

wherew119894is the 119894th row ofΩWith (10) the analysis dictionary

is learned from the clean signalX and the sparse constraint isenforced on ΩX In our proposed ADL algorithm howeverthe analysis dictionary is directly learned from the noisyobserved signal Y and thus can tolerate some sparsificationerrors as in [11] The results of the experiments demonstratethat our proposed approach is more robust

Note also that whenZ is sparse minimizing ΩY1can be

obtained by the minimization of Z minusΩY2119865 subject to the

sparsity constraint However both Z and Ω are unknownTo solve the problem we propose an iterative method toalternatively update the estimation of Z andΩ

21 Estimating Z Given the dictionary Ω we first considerthe optimization of Z only In this case the objective function(9) can be modified as

Z = arg minZ

(

1

2

Z minusΩY2119865+ 120582Z1) (11)

The first-order optimality condition of Z implies that

Z minusΩY + 120582 sign (Z) = 0 (12)

Therefore we have

119885119894119895=

(ΩY)119894119895 minus 120582 (ΩY)119894119895 gt 120582(ΩY)119894119895 + 120582 (ΩY)119894119895 lt minus1205820 otherwise

(13)

where 119894 = 1 2 119875 and 119895 = 1 2 119870 are the indices of thematrix elements It is well known that the above solution forZ is called soft thresholding Indeed the sparsity constraintZ1is only an approximation to Z

0 In order to promote

sparsity and to improve the approximation one could insteaduse the hard thresholding method [20] as an alternative thatis setting the smallest components of the vectors to be zeroswhile retaining the others

119885119894119895=

(ΩY)11989411989510038161003816100381610038161003816(ΩY)119894119895

10038161003816100381610038161003816gt 120582

0 otherwise(14)

As such the solution obtained using the constraint Z1will

be closer to that using Z0

22 Dictionary Learning For a given Z the objective func-tion (9) is nonconvex with respect to Ω similar to theobjective function (10) that is

Ω = arg minΩ

119891 (Ω)

= arg minΩ

(

1

2

Z minusΩY2119865+

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865)

(15)

The local minimum of the objective function (15) can befound by using a simple gradient descent method

Ω119905+1= Ω119905minus 120572nabla119891 (Ω) (16)

where 120572 is a step size and nabla119891(Ω) = minus(Z minusΩY)Y119879 + 120574(ΩΩ119879 minusI)Ω If the rows of Ω have different scales of norm wecannot directly use the hard thresholdingmethod to obtainZThe phenomenon is called scaling ambiguity Moreover theconstraint Ω119879Ω = I can enforce Ω to be of full column rankbut cannot avoid a subset of rows of Ω to be possibly zerosHence we normalize the rows of Ω to prevent the scalingambiguity and replace the zero rows if any by the normalizedrandom vectors that is

w119894=

w119894

1003817100381710038171003817w119894

10038171003817100381710038172

1003817100381710038171003817w119894

10038171003817100381710038172= 0

r otherwise(17)

4 The Scientific World Journal

where w119894is the 119894th row of Ω and r is a normalized random

vector There may be other methods to prevent this problemfor example by adding the normalization constraint to theobjective function [7] or by adding the mutual coherenceconstraint on the rows ofΩ to the objective function [12 14]however this is out of the scope of our work Although afterthe row normalization the columns of the dictionary areno longer close to the unit norm the rank of Ω will not bechanged

23 Convergence In the step of updating Z the algorithm tooptimize the objective function (11) is analytical Thus thealgorithm is guaranteed not to increase (11) and converges toa local minimum of (11) [21] Furthermore in the dictionaryupdate step Ω is updated by minimizing a fourth-orderfunction and thus amonotonic reduction in the cost function(15) is guaranteed

Our algorithm (called orthogonality constrained ADLwith iterative hard thresholding OIHT-ADL) to learn anal-ysis dictionary is summarized in Algorithm 1

3 Computer Simulations

To validate the proposed algorithm we perform two exper-iments In the first experiment we show the performanceof the proposed algorithm for synthetic dictionary recoveryproblems In the second experiment we consider the naturalimage denoising problems In these experimentsΩ

0isin 119877119875times119872

is the initial dictionary in which each row is orthogonal to arandom set of119872minus1 training data and is also normalized [10]For performance comparison the AK-SVD [10] NAAOLA[8] and LOST [12] algorithms are used as baselines

31 Experiments on Synthetic Data Following the work in[10] synthetic data are used to demonstrate the performanceof the proposed algorithm in recovering an underlyingdictionary Ω In this experiment we utilize the methods torecover a dictionary that is used to produce the set of trainingdata The convergence curves and the dictionary recoverypercentage are used to show their performance If min

119894(1 minus

|w119879119894w119895|) lt 001 a roww119879

119895in the true dictionaryΩ is regarded

as recovered where w119879119894is an atom of the trained dictionary

Ω isin 11987750times25 is generated with random Gaussian entries and

the dataset consists of 119870 = 50000 signals each residing in a4-dimensional subspace with both the noise-free setup andthe noise setup (120590 = 004 SNR = 25 dB) The parameters ofthe proposed algorithm are set empirically by experimentaltests and we choose the parameters as 120582 = 01 120574 = 10and 120572 = 10minus4 which give the best results in atom recoveryrate We have tested the parameters with different values inthe ranges of 10minus2 lt 120582 lt 10

2 10minus2 lt 120574 lt 102 and

10minus2lt 120572 lt 10

minus5 For the AK-SVD algorithm the dimensionof the subspace is set to be 119903 = 4 following the setting in [10]For the NAAOLA algorithm we set 120582 = 05 which is closeto the default value 120582 = 03 in [8] The parameters in theLOST algorithms are set as those in [11] that is 120572 = 10minus4120582 = 120578 = 120583 = 10

5 119901 = 20 and 119904 = 29 The convergence

curves of the error by the first term of (7) versus the iterationnumber are shown in Figure 1(a) It can be observed that thecompared algorithms take different numbers of iterations toconverge (the terms used tomeasure the rates of convergenceare slightly different in these algorithms because differentcost functions have been used) which are approximately200 100 1000 and 300 iterations for the OIHT-ADL AK-SVD NAAOLA and LOST algorithms respectively Thusthe recovery percentages of these algorithms are measuredat these different iteration numbers (to ensure that eachalgorithm converges to the stable solutions) It is observedfrom Figure 2 that 76 94 42 and 72 of the rows inthe true dictionary Ω are recovered for the noise-free caseand 72 86 12 and 68 for the noisy case respectivelyNote that the recovery rates for the LOST algorithm areobtained with random initialization With the DCT andidentitymatrix initialization the LOST algorithm can recover74 and 72 rows of the true dictionary Ω in noise-freeand noise cases respectively after 300 iterations Althoughit may not be necessary to recover the true dictionary inpractical applications the use of atom recovery rate allowsus to compare the proposed algorithm with the baselineADL algorithms as it has been widely used in these worksThe recovery rate by the NAAOLA algorithm is relativelylow mainly because it employs a uniformly normalized tightframe as a constraint on the dictionary and this constraint haslimited the possibleΩ to be learned The AK-SVD algorithmperforms the best in terms of the recovery rate that istaking fewer iterations to reach a similar recovery percentageHowever the running time in each iteration of the AK-SVDalgorithm is significantly higher than that in our proposedOIHT-ADL algorithm The total runtime of our algorithmfor 200 iterations is about 825 and 832 seconds for the noise-free and noise case respectively In contrast the AK-SVDalgorithm takes about 11544 or 10948 seconds respectivelyfor only 100 iterations (Computer OSWindows 7 CPU IntelCore i5-3210M 250GHz and RAM 4GB)This is becauseour algorithm does not need to estimate X in each iterationas opposed to the AK-SVD algorithm

32 Experiments on Natural Image Denoising In this exper-iment the training set of 20000 image patches each of7 times 7 pixels obtained from three images (Lena Houseand Peppers) that are commonly used in denoising [10]has been used for learning the dictionaries each of whichcorresponds to a particular noisy image Noise with differentlevel 120590 varying from 5 to 20 is added to these imagepatches The dictionary of size 63 times 49 is learned fromthe training data which are normalized to have zero meanFor fair comparison the dictionary size is the same as thatin [10] The dictionary with a bigger size may lead to asparser representation of the signal but may have a highercomputational cost Similar to Section 31 enough iterationsare performed for the tested algorithms to converge Theparameters for the OIHT-ADL algorithm are set the same asthose in Section 31 except that 120582 = 005 For the AK-SVDalgorithm the dimension of the subspace is set to be 119903 = 7following the setting in [10] For the NAAOLA algorithm

The Scientific World Journal 5

0 50 100 150 200800

1000

1200

1400

1600

1800

2000

Iteration number

05X

minusΩY

2 F

(a) OHIT-ADL

0 20 40 60 80 100001

002

003

004

005

006

007

008

Iteration number

XminusY

Fradic

MK

(b) AK-SVD

Noisy SNR = 25

0 500 1000 1500 2000538

539

54

541

542

543

544

545

546

Iteration number

Noise-free

log10(ΩY

1)

(c) NAAOLA

0 50 100 150 200 250 3000

200

400

600

800

1000

1200

1400

1600

Iteration number

Noise-free

05X

minusΩY

2 F

Noisy SNR = 25

(d) LOST

Figure 1 The convergence curves of the ADL algorithms for both the noise-free and noise case (where SNR = 25 dB)

Input Observed data Y isin 119877119872times119870 the initial dictionaryΩ0isin 119877119875times119872 120582 120574 120572 the number of iterations 119888

Output DictionaryΩInitialization SetΩ = Ω

0

For 119905 = 1 119888 do(i) FixingΩ computeΩY(ii) Update the analysis coefficients Z using (14)(iii) Fixing Z update the dictionaryΩ using (16)(iv) Normalize rows inΩ using (17)

End

Algorithm 1 OIHT-ADL

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 2: Research Article An Analysis Dictionary Learning Algorithm

2 The Scientific World Journal

be recovered by solving the following analysis model basedoptimization problem

x = arg min Ωx0 subject to 1003817100381710038171003817y minus x100381710038171003817

1003817

2

2le 120576 (5)

For the squared and invertible dictionary the synthesis andthe analysis models are identical withD = Ωminus1 [9] Howeverthe two models depart if we concentrate on the redundantcase (119875 gt 119872 and 119873 gt 119872) In the analysis model the signalx is described by the zero elements of z that is zero entriesin z that define a subspace which the signal x belongs toas opposed to the few nonzero entries of a in the synthesismodel

The performance of both the synthesis and analysismodels hinges on the representation of the signals with anappropriately chosen dictionary In the past decade a greatdeal of effort has been dedicated to learning the dictionary forthe synthesis model however the ADL problem has receivedless attention with only a few algorithms proposed recently[5 6 10 17ndash19] In these works the dictionary Ω is oftenlearned from the observed signals Y = [y

1 y2 y

119870] isin

119877119872times119870 measured in the presence of additive noise that is

Y = X + V (6)

where X = [x1 x2 x

119870] isin 119877

119872times119870 contains the originalsignals V = [k

1 k2 k

119870] isin 119877

119872times119870 is a Gaussian noisematrix and 119870 is the number of signals

Exploiting the fact that a row of the dictionary Ω isorthogonal to a subset of training signals X a sequentialminimal eigenvalue based ADL algorithm is proposed in [5]Once the subset is found the corresponding row of the dic-tionary can be updated with the eigenvector associated withthe smallest eigenvalue of the autocorrelation matrix of thesesignals However as 119875 increases so does the computationalcost of the method In [6ndash8] an ℓ

1-norm penalty function

is applied to Ωx and a projected subgradient algorithm(called NAAOLA) is proposed for analysis operator learningThese works employ a uniformly normalized tight frame asa constraint on the dictionary to avoid the trivial solution(eg Ω = 0) which however limits the range of possibleΩ to be learned In [10] the Analysis K-SVD (AK-SVD)algorithm is proposed for ADL By keeping Ω fixed theoptimal backward greedy algorithm (OBG) is employed toestimate a submatrix of Ω whose rows are then used todetermine a submatrix of Y and the eigenvector associatedwith the smallest eigenvalue of this submatrix is then used toupdateΩ A generalized analysismodel that is the transformmodel is proposed in the learning sparsifying transform(LST) algorithm [11] Unlike the analysismodel (2) the sparserepresentation of a signal in the LSTmodel is not constrainedto lie in the range space ofΩ Such a generalization allows thetransform model to accommodate a wider class of signals Aclosed-form solution to the LST problem is also developedin [13] The LST algorithm [11] is further extended to theovercomplete case leading to the LOST algorithm in [12]The transform K-SVD algorithm proposed recently in [14] isessentially a combination of the ideas in the LOST and AK-SVD algorithms

There are two problems that have been observed oralready studied in some of the ADL algorithms discussedabove The first one is associated with the computationalcost in the optimization process of the ADL algorithmsFor example in the AK-SVD algorithm X needs to beestimated before learning the dictionary and as a result theoptimization process becomes computationally demandingThe second one is associated with the trivial solutions to thedictionary which may lead to spurious sparse representa-tions To eliminate such trivial solutions extra constraints arerequired such as the full-rank constraint on Ω employed in[11ndash13] and the mutual coherence constraint on the rows ofΩin [12 14]

In this paper we propose a new optimization model andalgorithm attempting to provide alternative solutions to thetwo potential problems mentioned above More specificallysimilar in spirit to our recent work [17] we directly use theobserved data to compute the approximate analysis sparserepresentation of the original signals without having topreestimate X for learning the dictionary This leads to acomputationally very efficient algorithm Moreover differentfrom the LOST algorithm we enforce an orthogonalityconstraint onΩ which as shown later has led to an improvedlearning performance

The paper is organized as follows In Section 2 we discussthe novel ADL model and algorithm In Section 3 we showsome experimental results before concluding the paper inSection 4

2 The Proposed ADL Algorithm

In some algorithms discussed above such as [10] the opti-mization criterion is based on X In practice however Xis unknown and required to be estimated from Y Thisunfortunately results in a computationally very slow opti-mization process as shown in Section 31 To address thisissue we introduce a new model and algorithm for ADLwhere similar to [17] we use ΩY

1as an approximation to

ΩX1 which is further replaced by Z

1and then used as

a new constraint for the reconstruction term Z = ΩY Thisleads to the following model of ADL

min(12

Z minusΩY2119865+ 120582Z1) (7)

where 120582 gt 0 is a regularization parameter According tothe analysis model (2) Z would be an analysis coefficientmatrix only if its columns lie in the range space of the analysisdictionary Ω However Z=ΩY is not explicitly enforced in(7) and hence it is called the transform coefficient matrix asin [19] Using (7) the analysis dictionaryΩ is learned from Ywithout having to preestimate X as opposed to the AK-SVDalgorithm so the computational effort for estimating X fromY is exempted

With the model (7) however there is a trivial solutionthat is Ω = 0 Z = 0 To avoid such a solution additionalconstraints on Ω are required In [12] a full column rankconstraint has been imposed on Ω through the use of anegative log determinant of Ω119879Ω However the inverse of

The Scientific World Journal 3

Table 1 Image denoising results (PSNR in dB)

120590 Noisy in dB ADL method Lena House Peppers

5 3415

OIHT-ADL 3826 3808 3770NAAOLA 3736 3704 3593AK-SVD 3844 3920 3793LOST 3816 3850 3747

10 2813

OIHT-ADL 3502 3479 3389NAAOLA 3273 3261 3112AK-SVD 3485 3532 3382LOST 3475 3477 3361

15 2461

OIHT-ADL 3317 3305 3169NAAOLA 3087 3082 2922AK-SVD 3259 3298 3128LOST 3288 3282 3138

20 2211

OIHT-ADL 3183 3171 3013NAAOLA 2965 2922 2740AK-SVD 3138 3151 2976LOST 3164 3147 2984

Ω119879Ω needs to be computed in the conjugate gradientmethod

in [12] and the inverse of Ω119879Ω may affect the stability ofthe numerical computation when the initial of Ω119879Ω is illconditioned To address this problem in our work a functiondefined on Ω119879Ω = I is employed as a constraint term whichenforcesΩ to be a full column rank matrix based on the factthat the ranks of Ω and its corresponding Gram matrix areequal This leads to the following new optimization criterion

min(12

Z minusΩY2119865+ 120582Z1) subject to Ω119879Ω = I (8)

Using a Lagrangian multiplier 120574 the optimization problemcan be reformulated as

min(12

Z minusΩY2119865+ 120582Z1 +

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865) (9)

It is worth noting that our proposed objective function(9) is essentially different from the one used in [7] In [7] theobjective function is defined as follows

min(ΩX1 +120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865+

120582

4

sum

119894

1003817100381710038171003817100381710038171003817

w119879119894w119894minus

119872

119875

1003817100381710038171003817100381710038171003817

2

2

) (10)

wherew119894is the 119894th row ofΩWith (10) the analysis dictionary

is learned from the clean signalX and the sparse constraint isenforced on ΩX In our proposed ADL algorithm howeverthe analysis dictionary is directly learned from the noisyobserved signal Y and thus can tolerate some sparsificationerrors as in [11] The results of the experiments demonstratethat our proposed approach is more robust

Note also that whenZ is sparse minimizing ΩY1can be

obtained by the minimization of Z minusΩY2119865 subject to the

sparsity constraint However both Z and Ω are unknownTo solve the problem we propose an iterative method toalternatively update the estimation of Z andΩ

21 Estimating Z Given the dictionary Ω we first considerthe optimization of Z only In this case the objective function(9) can be modified as

Z = arg minZ

(

1

2

Z minusΩY2119865+ 120582Z1) (11)

The first-order optimality condition of Z implies that

Z minusΩY + 120582 sign (Z) = 0 (12)

Therefore we have

119885119894119895=

(ΩY)119894119895 minus 120582 (ΩY)119894119895 gt 120582(ΩY)119894119895 + 120582 (ΩY)119894119895 lt minus1205820 otherwise

(13)

where 119894 = 1 2 119875 and 119895 = 1 2 119870 are the indices of thematrix elements It is well known that the above solution forZ is called soft thresholding Indeed the sparsity constraintZ1is only an approximation to Z

0 In order to promote

sparsity and to improve the approximation one could insteaduse the hard thresholding method [20] as an alternative thatis setting the smallest components of the vectors to be zeroswhile retaining the others

119885119894119895=

(ΩY)11989411989510038161003816100381610038161003816(ΩY)119894119895

10038161003816100381610038161003816gt 120582

0 otherwise(14)

As such the solution obtained using the constraint Z1will

be closer to that using Z0

22 Dictionary Learning For a given Z the objective func-tion (9) is nonconvex with respect to Ω similar to theobjective function (10) that is

Ω = arg minΩ

119891 (Ω)

= arg minΩ

(

1

2

Z minusΩY2119865+

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865)

(15)

The local minimum of the objective function (15) can befound by using a simple gradient descent method

Ω119905+1= Ω119905minus 120572nabla119891 (Ω) (16)

where 120572 is a step size and nabla119891(Ω) = minus(Z minusΩY)Y119879 + 120574(ΩΩ119879 minusI)Ω If the rows of Ω have different scales of norm wecannot directly use the hard thresholdingmethod to obtainZThe phenomenon is called scaling ambiguity Moreover theconstraint Ω119879Ω = I can enforce Ω to be of full column rankbut cannot avoid a subset of rows of Ω to be possibly zerosHence we normalize the rows of Ω to prevent the scalingambiguity and replace the zero rows if any by the normalizedrandom vectors that is

w119894=

w119894

1003817100381710038171003817w119894

10038171003817100381710038172

1003817100381710038171003817w119894

10038171003817100381710038172= 0

r otherwise(17)

4 The Scientific World Journal

where w119894is the 119894th row of Ω and r is a normalized random

vector There may be other methods to prevent this problemfor example by adding the normalization constraint to theobjective function [7] or by adding the mutual coherenceconstraint on the rows ofΩ to the objective function [12 14]however this is out of the scope of our work Although afterthe row normalization the columns of the dictionary areno longer close to the unit norm the rank of Ω will not bechanged

23 Convergence In the step of updating Z the algorithm tooptimize the objective function (11) is analytical Thus thealgorithm is guaranteed not to increase (11) and converges toa local minimum of (11) [21] Furthermore in the dictionaryupdate step Ω is updated by minimizing a fourth-orderfunction and thus amonotonic reduction in the cost function(15) is guaranteed

Our algorithm (called orthogonality constrained ADLwith iterative hard thresholding OIHT-ADL) to learn anal-ysis dictionary is summarized in Algorithm 1

3 Computer Simulations

To validate the proposed algorithm we perform two exper-iments In the first experiment we show the performanceof the proposed algorithm for synthetic dictionary recoveryproblems In the second experiment we consider the naturalimage denoising problems In these experimentsΩ

0isin 119877119875times119872

is the initial dictionary in which each row is orthogonal to arandom set of119872minus1 training data and is also normalized [10]For performance comparison the AK-SVD [10] NAAOLA[8] and LOST [12] algorithms are used as baselines

31 Experiments on Synthetic Data Following the work in[10] synthetic data are used to demonstrate the performanceof the proposed algorithm in recovering an underlyingdictionary Ω In this experiment we utilize the methods torecover a dictionary that is used to produce the set of trainingdata The convergence curves and the dictionary recoverypercentage are used to show their performance If min

119894(1 minus

|w119879119894w119895|) lt 001 a roww119879

119895in the true dictionaryΩ is regarded

as recovered where w119879119894is an atom of the trained dictionary

Ω isin 11987750times25 is generated with random Gaussian entries and

the dataset consists of 119870 = 50000 signals each residing in a4-dimensional subspace with both the noise-free setup andthe noise setup (120590 = 004 SNR = 25 dB) The parameters ofthe proposed algorithm are set empirically by experimentaltests and we choose the parameters as 120582 = 01 120574 = 10and 120572 = 10minus4 which give the best results in atom recoveryrate We have tested the parameters with different values inthe ranges of 10minus2 lt 120582 lt 10

2 10minus2 lt 120574 lt 102 and

10minus2lt 120572 lt 10

minus5 For the AK-SVD algorithm the dimensionof the subspace is set to be 119903 = 4 following the setting in [10]For the NAAOLA algorithm we set 120582 = 05 which is closeto the default value 120582 = 03 in [8] The parameters in theLOST algorithms are set as those in [11] that is 120572 = 10minus4120582 = 120578 = 120583 = 10

5 119901 = 20 and 119904 = 29 The convergence

curves of the error by the first term of (7) versus the iterationnumber are shown in Figure 1(a) It can be observed that thecompared algorithms take different numbers of iterations toconverge (the terms used tomeasure the rates of convergenceare slightly different in these algorithms because differentcost functions have been used) which are approximately200 100 1000 and 300 iterations for the OIHT-ADL AK-SVD NAAOLA and LOST algorithms respectively Thusthe recovery percentages of these algorithms are measuredat these different iteration numbers (to ensure that eachalgorithm converges to the stable solutions) It is observedfrom Figure 2 that 76 94 42 and 72 of the rows inthe true dictionary Ω are recovered for the noise-free caseand 72 86 12 and 68 for the noisy case respectivelyNote that the recovery rates for the LOST algorithm areobtained with random initialization With the DCT andidentitymatrix initialization the LOST algorithm can recover74 and 72 rows of the true dictionary Ω in noise-freeand noise cases respectively after 300 iterations Althoughit may not be necessary to recover the true dictionary inpractical applications the use of atom recovery rate allowsus to compare the proposed algorithm with the baselineADL algorithms as it has been widely used in these worksThe recovery rate by the NAAOLA algorithm is relativelylow mainly because it employs a uniformly normalized tightframe as a constraint on the dictionary and this constraint haslimited the possibleΩ to be learned The AK-SVD algorithmperforms the best in terms of the recovery rate that istaking fewer iterations to reach a similar recovery percentageHowever the running time in each iteration of the AK-SVDalgorithm is significantly higher than that in our proposedOIHT-ADL algorithm The total runtime of our algorithmfor 200 iterations is about 825 and 832 seconds for the noise-free and noise case respectively In contrast the AK-SVDalgorithm takes about 11544 or 10948 seconds respectivelyfor only 100 iterations (Computer OSWindows 7 CPU IntelCore i5-3210M 250GHz and RAM 4GB)This is becauseour algorithm does not need to estimate X in each iterationas opposed to the AK-SVD algorithm

32 Experiments on Natural Image Denoising In this exper-iment the training set of 20000 image patches each of7 times 7 pixels obtained from three images (Lena Houseand Peppers) that are commonly used in denoising [10]has been used for learning the dictionaries each of whichcorresponds to a particular noisy image Noise with differentlevel 120590 varying from 5 to 20 is added to these imagepatches The dictionary of size 63 times 49 is learned fromthe training data which are normalized to have zero meanFor fair comparison the dictionary size is the same as thatin [10] The dictionary with a bigger size may lead to asparser representation of the signal but may have a highercomputational cost Similar to Section 31 enough iterationsare performed for the tested algorithms to converge Theparameters for the OIHT-ADL algorithm are set the same asthose in Section 31 except that 120582 = 005 For the AK-SVDalgorithm the dimension of the subspace is set to be 119903 = 7following the setting in [10] For the NAAOLA algorithm

The Scientific World Journal 5

0 50 100 150 200800

1000

1200

1400

1600

1800

2000

Iteration number

05X

minusΩY

2 F

(a) OHIT-ADL

0 20 40 60 80 100001

002

003

004

005

006

007

008

Iteration number

XminusY

Fradic

MK

(b) AK-SVD

Noisy SNR = 25

0 500 1000 1500 2000538

539

54

541

542

543

544

545

546

Iteration number

Noise-free

log10(ΩY

1)

(c) NAAOLA

0 50 100 150 200 250 3000

200

400

600

800

1000

1200

1400

1600

Iteration number

Noise-free

05X

minusΩY

2 F

Noisy SNR = 25

(d) LOST

Figure 1 The convergence curves of the ADL algorithms for both the noise-free and noise case (where SNR = 25 dB)

Input Observed data Y isin 119877119872times119870 the initial dictionaryΩ0isin 119877119875times119872 120582 120574 120572 the number of iterations 119888

Output DictionaryΩInitialization SetΩ = Ω

0

For 119905 = 1 119888 do(i) FixingΩ computeΩY(ii) Update the analysis coefficients Z using (14)(iii) Fixing Z update the dictionaryΩ using (16)(iv) Normalize rows inΩ using (17)

End

Algorithm 1 OIHT-ADL

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 3: Research Article An Analysis Dictionary Learning Algorithm

The Scientific World Journal 3

Table 1 Image denoising results (PSNR in dB)

120590 Noisy in dB ADL method Lena House Peppers

5 3415

OIHT-ADL 3826 3808 3770NAAOLA 3736 3704 3593AK-SVD 3844 3920 3793LOST 3816 3850 3747

10 2813

OIHT-ADL 3502 3479 3389NAAOLA 3273 3261 3112AK-SVD 3485 3532 3382LOST 3475 3477 3361

15 2461

OIHT-ADL 3317 3305 3169NAAOLA 3087 3082 2922AK-SVD 3259 3298 3128LOST 3288 3282 3138

20 2211

OIHT-ADL 3183 3171 3013NAAOLA 2965 2922 2740AK-SVD 3138 3151 2976LOST 3164 3147 2984

Ω119879Ω needs to be computed in the conjugate gradientmethod

in [12] and the inverse of Ω119879Ω may affect the stability ofthe numerical computation when the initial of Ω119879Ω is illconditioned To address this problem in our work a functiondefined on Ω119879Ω = I is employed as a constraint term whichenforcesΩ to be a full column rank matrix based on the factthat the ranks of Ω and its corresponding Gram matrix areequal This leads to the following new optimization criterion

min(12

Z minusΩY2119865+ 120582Z1) subject to Ω119879Ω = I (8)

Using a Lagrangian multiplier 120574 the optimization problemcan be reformulated as

min(12

Z minusΩY2119865+ 120582Z1 +

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865) (9)

It is worth noting that our proposed objective function(9) is essentially different from the one used in [7] In [7] theobjective function is defined as follows

min(ΩX1 +120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865+

120582

4

sum

119894

1003817100381710038171003817100381710038171003817

w119879119894w119894minus

119872

119875

1003817100381710038171003817100381710038171003817

2

2

) (10)

wherew119894is the 119894th row ofΩWith (10) the analysis dictionary

is learned from the clean signalX and the sparse constraint isenforced on ΩX In our proposed ADL algorithm howeverthe analysis dictionary is directly learned from the noisyobserved signal Y and thus can tolerate some sparsificationerrors as in [11] The results of the experiments demonstratethat our proposed approach is more robust

Note also that whenZ is sparse minimizing ΩY1can be

obtained by the minimization of Z minusΩY2119865 subject to the

sparsity constraint However both Z and Ω are unknownTo solve the problem we propose an iterative method toalternatively update the estimation of Z andΩ

21 Estimating Z Given the dictionary Ω we first considerthe optimization of Z only In this case the objective function(9) can be modified as

Z = arg minZ

(

1

2

Z minusΩY2119865+ 120582Z1) (11)

The first-order optimality condition of Z implies that

Z minusΩY + 120582 sign (Z) = 0 (12)

Therefore we have

119885119894119895=

(ΩY)119894119895 minus 120582 (ΩY)119894119895 gt 120582(ΩY)119894119895 + 120582 (ΩY)119894119895 lt minus1205820 otherwise

(13)

where 119894 = 1 2 119875 and 119895 = 1 2 119870 are the indices of thematrix elements It is well known that the above solution forZ is called soft thresholding Indeed the sparsity constraintZ1is only an approximation to Z

0 In order to promote

sparsity and to improve the approximation one could insteaduse the hard thresholding method [20] as an alternative thatis setting the smallest components of the vectors to be zeroswhile retaining the others

119885119894119895=

(ΩY)11989411989510038161003816100381610038161003816(ΩY)119894119895

10038161003816100381610038161003816gt 120582

0 otherwise(14)

As such the solution obtained using the constraint Z1will

be closer to that using Z0

22 Dictionary Learning For a given Z the objective func-tion (9) is nonconvex with respect to Ω similar to theobjective function (10) that is

Ω = arg minΩ

119891 (Ω)

= arg minΩ

(

1

2

Z minusΩY2119865+

120574

4

10038171003817100381710038171003817Ω119879Ω minus I1003817100381710038171003817

1003817

2

119865)

(15)

The local minimum of the objective function (15) can befound by using a simple gradient descent method

Ω119905+1= Ω119905minus 120572nabla119891 (Ω) (16)

where 120572 is a step size and nabla119891(Ω) = minus(Z minusΩY)Y119879 + 120574(ΩΩ119879 minusI)Ω If the rows of Ω have different scales of norm wecannot directly use the hard thresholdingmethod to obtainZThe phenomenon is called scaling ambiguity Moreover theconstraint Ω119879Ω = I can enforce Ω to be of full column rankbut cannot avoid a subset of rows of Ω to be possibly zerosHence we normalize the rows of Ω to prevent the scalingambiguity and replace the zero rows if any by the normalizedrandom vectors that is

w119894=

w119894

1003817100381710038171003817w119894

10038171003817100381710038172

1003817100381710038171003817w119894

10038171003817100381710038172= 0

r otherwise(17)

4 The Scientific World Journal

where w119894is the 119894th row of Ω and r is a normalized random

vector There may be other methods to prevent this problemfor example by adding the normalization constraint to theobjective function [7] or by adding the mutual coherenceconstraint on the rows ofΩ to the objective function [12 14]however this is out of the scope of our work Although afterthe row normalization the columns of the dictionary areno longer close to the unit norm the rank of Ω will not bechanged

23 Convergence In the step of updating Z the algorithm tooptimize the objective function (11) is analytical Thus thealgorithm is guaranteed not to increase (11) and converges toa local minimum of (11) [21] Furthermore in the dictionaryupdate step Ω is updated by minimizing a fourth-orderfunction and thus amonotonic reduction in the cost function(15) is guaranteed

Our algorithm (called orthogonality constrained ADLwith iterative hard thresholding OIHT-ADL) to learn anal-ysis dictionary is summarized in Algorithm 1

3 Computer Simulations

To validate the proposed algorithm we perform two exper-iments In the first experiment we show the performanceof the proposed algorithm for synthetic dictionary recoveryproblems In the second experiment we consider the naturalimage denoising problems In these experimentsΩ

0isin 119877119875times119872

is the initial dictionary in which each row is orthogonal to arandom set of119872minus1 training data and is also normalized [10]For performance comparison the AK-SVD [10] NAAOLA[8] and LOST [12] algorithms are used as baselines

31 Experiments on Synthetic Data Following the work in[10] synthetic data are used to demonstrate the performanceof the proposed algorithm in recovering an underlyingdictionary Ω In this experiment we utilize the methods torecover a dictionary that is used to produce the set of trainingdata The convergence curves and the dictionary recoverypercentage are used to show their performance If min

119894(1 minus

|w119879119894w119895|) lt 001 a roww119879

119895in the true dictionaryΩ is regarded

as recovered where w119879119894is an atom of the trained dictionary

Ω isin 11987750times25 is generated with random Gaussian entries and

the dataset consists of 119870 = 50000 signals each residing in a4-dimensional subspace with both the noise-free setup andthe noise setup (120590 = 004 SNR = 25 dB) The parameters ofthe proposed algorithm are set empirically by experimentaltests and we choose the parameters as 120582 = 01 120574 = 10and 120572 = 10minus4 which give the best results in atom recoveryrate We have tested the parameters with different values inthe ranges of 10minus2 lt 120582 lt 10

2 10minus2 lt 120574 lt 102 and

10minus2lt 120572 lt 10

minus5 For the AK-SVD algorithm the dimensionof the subspace is set to be 119903 = 4 following the setting in [10]For the NAAOLA algorithm we set 120582 = 05 which is closeto the default value 120582 = 03 in [8] The parameters in theLOST algorithms are set as those in [11] that is 120572 = 10minus4120582 = 120578 = 120583 = 10

5 119901 = 20 and 119904 = 29 The convergence

curves of the error by the first term of (7) versus the iterationnumber are shown in Figure 1(a) It can be observed that thecompared algorithms take different numbers of iterations toconverge (the terms used tomeasure the rates of convergenceare slightly different in these algorithms because differentcost functions have been used) which are approximately200 100 1000 and 300 iterations for the OIHT-ADL AK-SVD NAAOLA and LOST algorithms respectively Thusthe recovery percentages of these algorithms are measuredat these different iteration numbers (to ensure that eachalgorithm converges to the stable solutions) It is observedfrom Figure 2 that 76 94 42 and 72 of the rows inthe true dictionary Ω are recovered for the noise-free caseand 72 86 12 and 68 for the noisy case respectivelyNote that the recovery rates for the LOST algorithm areobtained with random initialization With the DCT andidentitymatrix initialization the LOST algorithm can recover74 and 72 rows of the true dictionary Ω in noise-freeand noise cases respectively after 300 iterations Althoughit may not be necessary to recover the true dictionary inpractical applications the use of atom recovery rate allowsus to compare the proposed algorithm with the baselineADL algorithms as it has been widely used in these worksThe recovery rate by the NAAOLA algorithm is relativelylow mainly because it employs a uniformly normalized tightframe as a constraint on the dictionary and this constraint haslimited the possibleΩ to be learned The AK-SVD algorithmperforms the best in terms of the recovery rate that istaking fewer iterations to reach a similar recovery percentageHowever the running time in each iteration of the AK-SVDalgorithm is significantly higher than that in our proposedOIHT-ADL algorithm The total runtime of our algorithmfor 200 iterations is about 825 and 832 seconds for the noise-free and noise case respectively In contrast the AK-SVDalgorithm takes about 11544 or 10948 seconds respectivelyfor only 100 iterations (Computer OSWindows 7 CPU IntelCore i5-3210M 250GHz and RAM 4GB)This is becauseour algorithm does not need to estimate X in each iterationas opposed to the AK-SVD algorithm

32 Experiments on Natural Image Denoising In this exper-iment the training set of 20000 image patches each of7 times 7 pixels obtained from three images (Lena Houseand Peppers) that are commonly used in denoising [10]has been used for learning the dictionaries each of whichcorresponds to a particular noisy image Noise with differentlevel 120590 varying from 5 to 20 is added to these imagepatches The dictionary of size 63 times 49 is learned fromthe training data which are normalized to have zero meanFor fair comparison the dictionary size is the same as thatin [10] The dictionary with a bigger size may lead to asparser representation of the signal but may have a highercomputational cost Similar to Section 31 enough iterationsare performed for the tested algorithms to converge Theparameters for the OIHT-ADL algorithm are set the same asthose in Section 31 except that 120582 = 005 For the AK-SVDalgorithm the dimension of the subspace is set to be 119903 = 7following the setting in [10] For the NAAOLA algorithm

The Scientific World Journal 5

0 50 100 150 200800

1000

1200

1400

1600

1800

2000

Iteration number

05X

minusΩY

2 F

(a) OHIT-ADL

0 20 40 60 80 100001

002

003

004

005

006

007

008

Iteration number

XminusY

Fradic

MK

(b) AK-SVD

Noisy SNR = 25

0 500 1000 1500 2000538

539

54

541

542

543

544

545

546

Iteration number

Noise-free

log10(ΩY

1)

(c) NAAOLA

0 50 100 150 200 250 3000

200

400

600

800

1000

1200

1400

1600

Iteration number

Noise-free

05X

minusΩY

2 F

Noisy SNR = 25

(d) LOST

Figure 1 The convergence curves of the ADL algorithms for both the noise-free and noise case (where SNR = 25 dB)

Input Observed data Y isin 119877119872times119870 the initial dictionaryΩ0isin 119877119875times119872 120582 120574 120572 the number of iterations 119888

Output DictionaryΩInitialization SetΩ = Ω

0

For 119905 = 1 119888 do(i) FixingΩ computeΩY(ii) Update the analysis coefficients Z using (14)(iii) Fixing Z update the dictionaryΩ using (16)(iv) Normalize rows inΩ using (17)

End

Algorithm 1 OIHT-ADL

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 4: Research Article An Analysis Dictionary Learning Algorithm

4 The Scientific World Journal

where w119894is the 119894th row of Ω and r is a normalized random

vector There may be other methods to prevent this problemfor example by adding the normalization constraint to theobjective function [7] or by adding the mutual coherenceconstraint on the rows ofΩ to the objective function [12 14]however this is out of the scope of our work Although afterthe row normalization the columns of the dictionary areno longer close to the unit norm the rank of Ω will not bechanged

23 Convergence In the step of updating Z the algorithm tooptimize the objective function (11) is analytical Thus thealgorithm is guaranteed not to increase (11) and converges toa local minimum of (11) [21] Furthermore in the dictionaryupdate step Ω is updated by minimizing a fourth-orderfunction and thus amonotonic reduction in the cost function(15) is guaranteed

Our algorithm (called orthogonality constrained ADLwith iterative hard thresholding OIHT-ADL) to learn anal-ysis dictionary is summarized in Algorithm 1

3 Computer Simulations

To validate the proposed algorithm we perform two exper-iments In the first experiment we show the performanceof the proposed algorithm for synthetic dictionary recoveryproblems In the second experiment we consider the naturalimage denoising problems In these experimentsΩ

0isin 119877119875times119872

is the initial dictionary in which each row is orthogonal to arandom set of119872minus1 training data and is also normalized [10]For performance comparison the AK-SVD [10] NAAOLA[8] and LOST [12] algorithms are used as baselines

31 Experiments on Synthetic Data Following the work in[10] synthetic data are used to demonstrate the performanceof the proposed algorithm in recovering an underlyingdictionary Ω In this experiment we utilize the methods torecover a dictionary that is used to produce the set of trainingdata The convergence curves and the dictionary recoverypercentage are used to show their performance If min

119894(1 minus

|w119879119894w119895|) lt 001 a roww119879

119895in the true dictionaryΩ is regarded

as recovered where w119879119894is an atom of the trained dictionary

Ω isin 11987750times25 is generated with random Gaussian entries and

the dataset consists of 119870 = 50000 signals each residing in a4-dimensional subspace with both the noise-free setup andthe noise setup (120590 = 004 SNR = 25 dB) The parameters ofthe proposed algorithm are set empirically by experimentaltests and we choose the parameters as 120582 = 01 120574 = 10and 120572 = 10minus4 which give the best results in atom recoveryrate We have tested the parameters with different values inthe ranges of 10minus2 lt 120582 lt 10

2 10minus2 lt 120574 lt 102 and

10minus2lt 120572 lt 10

minus5 For the AK-SVD algorithm the dimensionof the subspace is set to be 119903 = 4 following the setting in [10]For the NAAOLA algorithm we set 120582 = 05 which is closeto the default value 120582 = 03 in [8] The parameters in theLOST algorithms are set as those in [11] that is 120572 = 10minus4120582 = 120578 = 120583 = 10

5 119901 = 20 and 119904 = 29 The convergence

curves of the error by the first term of (7) versus the iterationnumber are shown in Figure 1(a) It can be observed that thecompared algorithms take different numbers of iterations toconverge (the terms used tomeasure the rates of convergenceare slightly different in these algorithms because differentcost functions have been used) which are approximately200 100 1000 and 300 iterations for the OIHT-ADL AK-SVD NAAOLA and LOST algorithms respectively Thusthe recovery percentages of these algorithms are measuredat these different iteration numbers (to ensure that eachalgorithm converges to the stable solutions) It is observedfrom Figure 2 that 76 94 42 and 72 of the rows inthe true dictionary Ω are recovered for the noise-free caseand 72 86 12 and 68 for the noisy case respectivelyNote that the recovery rates for the LOST algorithm areobtained with random initialization With the DCT andidentitymatrix initialization the LOST algorithm can recover74 and 72 rows of the true dictionary Ω in noise-freeand noise cases respectively after 300 iterations Althoughit may not be necessary to recover the true dictionary inpractical applications the use of atom recovery rate allowsus to compare the proposed algorithm with the baselineADL algorithms as it has been widely used in these worksThe recovery rate by the NAAOLA algorithm is relativelylow mainly because it employs a uniformly normalized tightframe as a constraint on the dictionary and this constraint haslimited the possibleΩ to be learned The AK-SVD algorithmperforms the best in terms of the recovery rate that istaking fewer iterations to reach a similar recovery percentageHowever the running time in each iteration of the AK-SVDalgorithm is significantly higher than that in our proposedOIHT-ADL algorithm The total runtime of our algorithmfor 200 iterations is about 825 and 832 seconds for the noise-free and noise case respectively In contrast the AK-SVDalgorithm takes about 11544 or 10948 seconds respectivelyfor only 100 iterations (Computer OSWindows 7 CPU IntelCore i5-3210M 250GHz and RAM 4GB)This is becauseour algorithm does not need to estimate X in each iterationas opposed to the AK-SVD algorithm

32 Experiments on Natural Image Denoising In this exper-iment the training set of 20000 image patches each of7 times 7 pixels obtained from three images (Lena Houseand Peppers) that are commonly used in denoising [10]has been used for learning the dictionaries each of whichcorresponds to a particular noisy image Noise with differentlevel 120590 varying from 5 to 20 is added to these imagepatches The dictionary of size 63 times 49 is learned fromthe training data which are normalized to have zero meanFor fair comparison the dictionary size is the same as thatin [10] The dictionary with a bigger size may lead to asparser representation of the signal but may have a highercomputational cost Similar to Section 31 enough iterationsare performed for the tested algorithms to converge Theparameters for the OIHT-ADL algorithm are set the same asthose in Section 31 except that 120582 = 005 For the AK-SVDalgorithm the dimension of the subspace is set to be 119903 = 7following the setting in [10] For the NAAOLA algorithm

The Scientific World Journal 5

0 50 100 150 200800

1000

1200

1400

1600

1800

2000

Iteration number

05X

minusΩY

2 F

(a) OHIT-ADL

0 20 40 60 80 100001

002

003

004

005

006

007

008

Iteration number

XminusY

Fradic

MK

(b) AK-SVD

Noisy SNR = 25

0 500 1000 1500 2000538

539

54

541

542

543

544

545

546

Iteration number

Noise-free

log10(ΩY

1)

(c) NAAOLA

0 50 100 150 200 250 3000

200

400

600

800

1000

1200

1400

1600

Iteration number

Noise-free

05X

minusΩY

2 F

Noisy SNR = 25

(d) LOST

Figure 1 The convergence curves of the ADL algorithms for both the noise-free and noise case (where SNR = 25 dB)

Input Observed data Y isin 119877119872times119870 the initial dictionaryΩ0isin 119877119875times119872 120582 120574 120572 the number of iterations 119888

Output DictionaryΩInitialization SetΩ = Ω

0

For 119905 = 1 119888 do(i) FixingΩ computeΩY(ii) Update the analysis coefficients Z using (14)(iii) Fixing Z update the dictionaryΩ using (16)(iv) Normalize rows inΩ using (17)

End

Algorithm 1 OIHT-ADL

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 5: Research Article An Analysis Dictionary Learning Algorithm

The Scientific World Journal 5

0 50 100 150 200800

1000

1200

1400

1600

1800

2000

Iteration number

05X

minusΩY

2 F

(a) OHIT-ADL

0 20 40 60 80 100001

002

003

004

005

006

007

008

Iteration number

XminusY

Fradic

MK

(b) AK-SVD

Noisy SNR = 25

0 500 1000 1500 2000538

539

54

541

542

543

544

545

546

Iteration number

Noise-free

log10(ΩY

1)

(c) NAAOLA

0 50 100 150 200 250 3000

200

400

600

800

1000

1200

1400

1600

Iteration number

Noise-free

05X

minusΩY

2 F

Noisy SNR = 25

(d) LOST

Figure 1 The convergence curves of the ADL algorithms for both the noise-free and noise case (where SNR = 25 dB)

Input Observed data Y isin 119877119872times119870 the initial dictionaryΩ0isin 119877119875times119872 120582 120574 120572 the number of iterations 119888

Output DictionaryΩInitialization SetΩ = Ω

0

For 119905 = 1 119888 do(i) FixingΩ computeΩY(ii) Update the analysis coefficients Z using (14)(iii) Fixing Z update the dictionaryΩ using (16)(iv) Normalize rows inΩ using (17)

End

Algorithm 1 OIHT-ADL

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 6: Research Article An Analysis Dictionary Learning Algorithm

6 The Scientific World Journal

0 50 100 150 2000

20

40

60

80

100

Iteration number

The r

ecov

ered

row

s (

)

(a) OHIT-ADL

0

20

40

60

80

100

0 20 40 60 80 100

Iteration number

The r

ecov

ered

row

s (

)

(b) AK-SVD

0 500 1000 1500 20000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(c) NAAOLA

0 50 100 150 200 250 3000

20

40

60

80

100

Iteration number

Noise-free

The r

ecov

ered

row

s (

)

Noisy SNR = 25

(d) LOST

Figure 2 The recovery percentage curves of the ADL algorithms

we set 120582 = 3 (120590 = 5) 120582 = 1 (120590 = 10 15) and 120582 = 05(120590 = 20) For the LOST algorithm the parameter values areset as 120572 = 10minus11 120582 = 120578 = 120583 = 105 119901 = 20 and 119904 = 21The parameters for the NAAOLA algorithm and the LOSTalgorithmare chosen empiricallyThe examples of the learneddictionaries are shown from the top to the bottom row inFigure 3 where each atom is shown as a 7 times 7 pixel imagepatch The atoms in the dictionaries learned by the AK-SVDalgorithm are sorted by the number of training patches whichare orthogonal to the corresponding atoms and thereforethe atoms learned by the AK-SVD algorithm appear to bemore structured Sorting is however not used in our proposedalgorithmThenweuse the learneddictionary to denoise eachoverlapping patch extracted from the noisy image Each patchis reconstructed individually and finally the entire image isformed by averaging over the overlapping regions We usethe OBG algorithm [10] to recover each patch image For

fair comparison we do not use the specific image denoisingmethod designed for the corresponding ADL algorithm suchas the one in [12] The denoising performance is evaluatedby the peak signal to noise ratio (PSNR) defined as PSNR =10 log

10(2552 KMsum119870

119894=1sum119872

119895=1(119909119894119895minus 119909119894119895)2) (dB) where 119909

119894119895and

119909119894119895are the 119894119895th pixel value in noisy and denoised images

respectivelyThe results averaged over 10 trials are presentedinTable 1 fromwhichwe can observe that the performance ofthe OIHT-ADL algorithm is generally better than that of theNAAOLA algorithm and the LOST algorithm It is also betterthan that of the AK-SVD algorithm when the noise level isincreased

4 Conclusion

Wehave presented a novel optimizationmodel and algorithmfor ADL where we learn the dictionary directly from the

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 7: Research Article An Analysis Dictionary Learning Algorithm

The Scientific World Journal 7

(a) Lena (b) House (c) Peppers

(d) Lena (e) House (f) Peppers

(g) Lena (h) House (i) Peppers

(j) Lena (k) House (l) Peppers

Figure 3 The learned dictionaries of size 63 times 49 by using the OIHT-ADL AK-SVD NAAOLA and LOST algorithms on the three imageswith noise level 120590 = 5 The results of the OIHT-ADL are shown in the top row in (a) Lena (b) House and (c) Peppers followed by theAK-SVD NAAOLA and LOST in the remaining rows

observed noisy dataWe have also enforced the orthogonalityconstraint on the optimization criterion to remove the trivialsolutions induced by a null dictionary matrix The proposedmethod offers computational advantage over the AK-SVDalgorithm and also provides an alternative to the LOST

algorithm in avoiding the trivial solutions The experimentsperformed have shown its competitive performance as com-pared to the three state-of-the-art algorithms the AK-SVDNAAOLA and LOST algorithms respectively The proposedalgorithm is easy to implement and computationally efficient

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 8: Research Article An Analysis Dictionary Learning Algorithm

8 The Scientific World Journal

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (Grants nos 61162014 and 61210306074)and the Natural Science Foundation of Jiangxi (Grant no20122BAB201025) andwas supported in part by the Engineer-ing and Physical Sciences Research Council (EPSRC) of theUK (Grant no EPK0143071)

References

[1] K Engan S O Aase and J H Husoy ldquoMethod of optimaldirections for frame designrdquo in IEEE International Conferenceon Acoustics Speech and Signal Processing vol 5 pp 2443ndash2446 1999

[2] M Aharon M Elad and A Bruckstein ldquoK-svd an algorithmfor designing overcomplete dictionaries for sparse representa-tionrdquo IEEE Transactions on Signal Processing vol 54 no 11 pp4311ndash4322 2006

[3] K Skretting and K Engan ldquoRecursive least squares dictionarylearning algorithmrdquo IEEE Transactions on Signal Processing vol58 no 4 pp 2121ndash2130 2010

[4] S Nam M E Davies M Elad and R Gribonval ldquoCosparseanalysis modelingmdashuniqueness and algorithmsrdquo in Proceedingsof the IEEE International Conference on Acoustics Speech andSignal Processing (ICASSP 11) pp 5804ndash5807 Prague CzechRepublic May 2011

[5] B Ophir M Elad N Bertin and M D Plumbley ldquoSequen-tial minimal eigenvalues an approach to analysis dictionarylearningrdquo in Proceedings of the 9th European Signal ProcessingConference pp 1465ndash1469 2011

[6] M Yaghoobi S Nam R Gribonval and M E Davies ldquoCon-strained overcomplete analysis operator learning for cosparsesignal modellingrdquo IEEE Transactions on Signal Processing vol61 no 9 pp 2341ndash2355 2013

[7] M Yaghoobi and M E Davies ldquoRelaxed analysis operatorlearningrdquo in Proceedings of the NIPS Workshop on AnalysisOperator Learning vs Dictionary Learning Fraternal Twins inSparse Modeling 2012

[8] M Yaghoobi S Nam R Gribonval and M E Davies ldquoNoiseaware analysis operator learning for approximately cosparsesignalsrdquo in Proceedings of the IEEE International Conference onAcoustics Speech and Signal Processing pp 5409ndash5412 2012

[9] M Elad P Milanfar and R Rubinstein ldquoAnalysis versussynthesis in signal priorsrdquo Inverse Problems vol 23 no 3 pp947ndash968 2007

[10] R Rubinstein T Peleg and M Elad ldquoAnalysis K-SVD adictionary-learning algorithm for the analysis sparse modelrdquoIEEE Transactions on Signal Processing vol 61 no 3 pp 661ndash677 2013

[11] S Ravishankar and Y Bresler ldquoLearning sparsifying transformsfor image processingrdquo in Proceedings of the IEEE InternationalConference on Image Processing pp 681ndash684 2012

[12] S Ravishankar and Y Bresler ldquoLearning overcomplete sparsi-fying transforms for signal processingrdquo in Proceedings of theIEEE International Conference on Acoustics Speech and Signal

Processing (ICASSP 13) pp 3088ndash3092 Vancouver CanadaMay 2013

[13] S Ravishankar and Y Bresler ldquoClosed-form solutions withinsparsifying transform learningrdquo inProceedings of the IEEE Inter-national Conference on Acoustics Speech and Signal Processingpp 5378ndash5382 2013

[14] E M Eksioglu and O Bayir ldquoK-svd meets transform learningtransform k-svdrdquo IEEE Signal Processing Letters vol 21 no 3pp 347ndash351 2014

[15] E J Candes J Romberg and T Tao ldquoRobust uncertaintyprinciples exact signal reconstruction from highly incompletefrequency informationrdquo IEEE Transactions on InformationThe-ory vol 52 no 2 pp 489ndash509 2006

[16] E J Candes and T Tao ldquoDecoding by linear programmingrdquoIEEE Transactions on Information Theory vol 51 no 12 pp4203ndash4215 2005

[17] Y Zhang H Wang T Yu and W Wang ldquoSubsetpursuit for analysis dictionary learningrdquo in Proceedingsof the European Signal Processing Conference 2013httppersonaleesurreyacukPersonalWWangpapersZha-ngWYW EUSIPCO 2013pdf

[18] Y Zhang W Wang H Wang and S Sanei ldquoKplane clusteringalgorithm for analysis dictionary learningrdquo in Proceedings of theIEEE International Workshop on Machine Learning for SignalProcessing pp 1ndash4 2013

[19] S Ravishankar and Y Bresler ldquoLearning sparsifying trans-formsrdquo IEEE Transactions on Signal Processing vol 61 no 5 pp1072ndash1086 2013

[20] M Elad M A T Figueiredo and Y Ma ldquoOn the role ofsparse and redundant representations in image processingrdquoProceedings of the IEEE vol 98 no 6 pp 972ndash982 2010

[21] T Blumensath and M E Davies ldquoIterative thresholding forsparse approximationsrdquo The Journal of Fourier Analysis andApplications vol 14 no 5-6 pp 629ndash654 2008

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 9: Research Article An Analysis Dictionary Learning Algorithm

International Journal of

AerospaceEngineeringHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal ofEngineeringVolume 2014

Submit your manuscripts athttpwwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of