11
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple Description Coding With Prediction Compensation Guoqian Sun, Upul Samarawickrama, Student Member, IEEE, Jie Liang, Member, IEEE, Chao Tian, Member, IEEE, Chengjie Tu, Member, IEEE, and Trac D. Tran, Senior Member, IEEE Abstract—A new multiple description coding paradigm is pro- posed by combining the time-domain lapped transform, block level source splitting, linear prediction, and prediction residual encoding. The method provides effective redundancy control and fully utilizes the source correlation. The joint optimization of all system components and the asymptotic performance analysis are presented. Image coding results demonstrate the superior perfor- mance of the proposed method, especially at low redundancies. Index Terms—Image coding, lapped transform, linear estima- tion, multiple description coding (MDC), rate distortion theory. I. INTRODUCTION M ULTIPLE DESCRIPTION coding (MDC) [1] is an at- tractive technique of combating transmission errors, in which several compressed bit streams (descriptions) are gen- erated and can be transmitted via different paths. The descrip- tions are designed such that the reconstruction quality degrades gracefully when some of them are lost, making MDC suitable for transmitting multimedia signals, where some transmission errors can be tolerated. A central issue in practical MDC designs is how to introduce a controlled amount of redundancy into the descriptions which the decoder can conveniently utilize, and many methods have been proposed. In [2], a multiple description scalar quantizer (MDSQ) is developed by using a central quantizer and an index assignment, which generates two side quantizers such that each of them alone produces an acceptable side distortion, whereas their combination yields the finer central quantizer. The MDSQ is asymptotically near optimal [3], and has been employed in, Manuscript received February 19, 2007; revised December 03, 2008. First published March 24, 2009; current version published April 10, 2009. This work was supported in part by the Natural Sciences and Engineering Research Council (NSERC) of Canada under Grants RGPIN312262-05, EQPEQ330976-2006, and STPGP350416-07. The associate editor coordi- nating the review of this manuscript and approving it for publication was Dr. Antonio Ortega. G. Sun was with Simon Fraser University, Burnaby, BC V5A 1S6 Canada. He is now with Cisco Systems, Inc., Vancouver, BC V7X 1J1 Canada (e-mail: [email protected]). U. Samarawickrama and J. Liang are with the Simon Fraser University, Burnaby, BC V5A 1S6 Canada (e-mail: [email protected]; [email protected]). C. Tian is with AT&T Labs-Research, Florham Park, NJ 07932 USA (e-mail: [email protected]). C. Tu is with Microsoft Corporation, Redmond, WA 98052 USA (e-mail: [email protected]). T. D. Tran is with the Electrical and Computer Engineering Department, The Johns Hopkins University, Baltimore, MD 21218 USA (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TIP.2009.2013068 e.g., [4] and [5], after the DCT or wavelet transform. However, the MDSQ index assignment is difficult to design and imple- ment, and its redundancy is not easy to adjust. In [6], a modified MDSQ (MMDSQ) with the same asymp- totical performance as the MDSQ is developed, in which two staggered scalar quantizers are used to generate the first layer of each description. Another scalar quantizer is used to further par- tition the joint bins of the first-layer quantizers, and its output is split into the two descriptions. The MMDSQ avoids the index assignment and can easily adjust the redundancy. It also outper- forms other MDSQ-based methods in MD image coding. How- ever, both MDSQ and MMDSQ do not perform well at low re- dundancy regime, which is a desired property of good MDC schemes [7, pp. 365]. Another family of MDC schemes is based on the source split- ting approach pioneered by Jayant in [8] and [9], where a signal is split into even and odd samples, and DPCM is used to encoded each description. If one description is lost, the missing data are predicted from their neighbors in the other description, using the source correlation. However, the prediction errors of the missing data are tied to the source correlation, which cannot be con- trolled. In [10], DPCM is used before splitting, and the predic- tion in the DPCM is designed to preserve some source correla- tions. Therefore, the redundancy between the descriptions can be adjusted to some extent. Although the method reduces the interdescription prediction error, the remaining error still limits the side decoder performance, especially at high rates. In [11], the transform coefficients are split into two parts. Each part is quantized into one description. To introduce re- dundancy, each description also includes a coarsely quantized version of the other part, which helps the decoding when the other description is lost. The optimal redundancy rate alloca- tion is studied. A similar approach is developed in [12] using the SPIHT algorithm. Recently, this method is applied to the JPEG 2000 framework in [13] under the name of RD-MDC, in which each JPEG 2000 code-block is encoded at two rates, one in each description, and the rate allocation is determined by Lagrangian optimization. However, to get balanced descrip- tions and optimal performance, the RD-MDC needs to classify all code-blocks into two subsets, such that any code-block in one subset has similar characteristics to another code-block in the other subset. This procedure is quite time-consuming. In ad- dition, the side distortion of the RD-MDC at low redundancies is not satisfactory, as shown in Section V. The pairwise correlating transform (PCT) [7] represents an- other method of introducing redundancy, where a set of 2 2 correlating transforms is applied to the uncorrelated coefficients 1057-7149/$25.00 © 2009 IEEE Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037

Multiple Description Coding WithPrediction Compensation

Guoqian Sun, Upul Samarawickrama, Student Member, IEEE, Jie Liang, Member, IEEE, Chao Tian, Member, IEEE,Chengjie Tu, Member, IEEE, and Trac D. Tran, Senior Member, IEEE

Abstract—A new multiple description coding paradigm is pro-posed by combining the time-domain lapped transform, blocklevel source splitting, linear prediction, and prediction residualencoding. The method provides effective redundancy control andfully utilizes the source correlation. The joint optimization of allsystem components and the asymptotic performance analysis arepresented. Image coding results demonstrate the superior perfor-mance of the proposed method, especially at low redundancies.

Index Terms—Image coding, lapped transform, linear estima-tion, multiple description coding (MDC), rate distortion theory.

I. INTRODUCTION

M ULTIPLE DESCRIPTION coding (MDC) [1] is an at-tractive technique of combating transmission errors, in

which several compressed bit streams (descriptions) are gen-erated and can be transmitted via different paths. The descrip-tions are designed such that the reconstruction quality degradesgracefully when some of them are lost, making MDC suitablefor transmitting multimedia signals, where some transmissionerrors can be tolerated.

A central issue in practical MDC designs is how to introducea controlled amount of redundancy into the descriptions whichthe decoder can conveniently utilize, and many methods havebeen proposed. In [2], a multiple description scalar quantizer(MDSQ) is developed by using a central quantizer and an indexassignment, which generates two side quantizers such that eachof them alone produces an acceptable side distortion, whereastheir combination yields the finer central quantizer. The MDSQis asymptotically near optimal [3], and has been employed in,

Manuscript received February 19, 2007; revised December 03, 2008.First published March 24, 2009; current version published April 10, 2009.This work was supported in part by the Natural Sciences and EngineeringResearch Council (NSERC) of Canada under Grants RGPIN312262-05,EQPEQ330976-2006, and STPGP350416-07. The associate editor coordi-nating the review of this manuscript and approving it for publication was Dr.Antonio Ortega.

G. Sun was with Simon Fraser University, Burnaby, BC V5A 1S6 Canada.He is now with Cisco Systems, Inc., Vancouver, BC V7X 1J1 Canada (e-mail:[email protected]).

U. Samarawickrama and J. Liang are with the Simon Fraser University,Burnaby, BC V5A 1S6 Canada (e-mail: [email protected]; [email protected]).

C. Tian is with AT&T Labs-Research, Florham Park, NJ 07932 USA (e-mail:[email protected]).

C. Tu is with Microsoft Corporation, Redmond, WA 98052 USA (e-mail:[email protected]).

T. D. Tran is with the Electrical and Computer Engineering Department, TheJohns Hopkins University, Baltimore, MD 21218 USA (e-mail: [email protected]).

Color versions of one or more of the figures in this paper are available onlineat http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TIP.2009.2013068

e.g., [4] and [5], after the DCT or wavelet transform. However,the MDSQ index assignment is difficult to design and imple-ment, and its redundancy is not easy to adjust.

In [6], a modified MDSQ (MMDSQ) with the same asymp-totical performance as the MDSQ is developed, in which twostaggered scalar quantizers are used to generate the first layer ofeach description. Another scalar quantizer is used to further par-tition the joint bins of the first-layer quantizers, and its output issplit into the two descriptions. The MMDSQ avoids the indexassignment and can easily adjust the redundancy. It also outper-forms other MDSQ-based methods in MD image coding. How-ever, both MDSQ and MMDSQ do not perform well at low re-dundancy regime, which is a desired property of good MDCschemes [7, pp. 365].

Another family of MDC schemes is based on the source split-ting approach pioneered by Jayant in [8] and [9], where a signalis split into even and odd samples, and DPCM is used to encodedeach description. If one description is lost, the missing data arepredicted from their neighbors in the other description, using thesource correlation. However, the prediction errors of the missingdata are tied to the source correlation, which cannot be con-trolled. In [10], DPCM is used before splitting, and the predic-tion in the DPCM is designed to preserve some source correla-tions. Therefore, the redundancy between the descriptions canbe adjusted to some extent. Although the method reduces theinterdescription prediction error, the remaining error still limitsthe side decoder performance, especially at high rates.

In [11], the transform coefficients are split into two parts.Each part is quantized into one description. To introduce re-dundancy, each description also includes a coarsely quantizedversion of the other part, which helps the decoding when theother description is lost. The optimal redundancy rate alloca-tion is studied. A similar approach is developed in [12] usingthe SPIHT algorithm. Recently, this method is applied to theJPEG 2000 framework in [13] under the name of RD-MDC,in which each JPEG 2000 code-block is encoded at two rates,one in each description, and the rate allocation is determinedby Lagrangian optimization. However, to get balanced descrip-tions and optimal performance, the RD-MDC needs to classifyall code-blocks into two subsets, such that any code-block inone subset has similar characteristics to another code-block inthe other subset. This procedure is quite time-consuming. In ad-dition, the side distortion of the RD-MDC at low redundanciesis not satisfactory, as shown in Section V.

The pairwise correlating transform (PCT) [7] represents an-other method of introducing redundancy, where a set of 2 2correlating transforms is applied to the uncorrelated coefficients

1057-7149/$25.00 © 2009 IEEE

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 2: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

1038 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009

after the DCT. The outputs of each PCT are split into two de-scriptions. If one coefficient is lost, it is estimated from its coun-terpart in the other description.

The PCT framework has some inherent drawbacks. First ofall, although the PCT has good low redundancy performance intheory, its practical application could not fully achieve this, be-cause the PCT can only be applied to coefficients with large vari-ances relative to the quantization error [7]. Other coefficientsare directly split into the two descriptions. In the side decoder,these low-variance coefficients are simply estimated as zero,which limits the side decoder performance at low redundancies.Second, similar to [8], the PCT does not perform well at high re-dundancies because of the prediction residual [14]. In this case,the decoded image of the side decoder in [7] can be 2 dB lowerthan the MDSQ.

Third, the PCT system only uses the correlation it inserts be-tween the two parts of a block, but does not exploit the rich cor-relation among neighboring blocks. Finally, the practical imple-mentation of the PCT system is not easy. Given coefficients,the method needs to pair the two coefficients with the th and

th largest variances, and the optimal PCTs depend onthe coefficient variances. To estimate these variances, in [7] allimage blocks are classified into four classes. Coefficient vari-ances of each class are then calculated and sorted for PCT de-signs. In addition, existing entropy coding for single descriptioncoding cannot be used for the PCT outputs due to different sta-tistics and block sizes.

A generalized PCT (GPCT) is proposed in [14], which is ahybrid method. At low redundancies, it is simply the PCT. Athigh redundancies, in addition to the PCT, each description alsoencodes the prediction residual of the other half of PCT outputs.It is shown in [14] that the GPCT side distortion is 3 dB awayfrom that of the MDSQ and MMDSQ. However, the GPCT onlyimproves the PCT at high redundancies, but does not solve otherproblems of the PCT. In addition, there has not been any prac-tical application using the GPCT.

As a summary, the methods in [2], [5]–[7], and [11]–[14]offer good redundancy control, but do not fully exploit thesource correlation. Although the schemes in [8] and [9] canachieve good performance at low redundancies by utilizing thesource correlation, they cannot adjust the redundancy of theMDC system. The algorithm in [10] uses the source correlationto adjust the redundancy, but suffers from the prediction error.Therefore, the goal of this paper is to develop an MDC schemethat can simultaneously achieve the following properties: pro-viding effective control of the redundancy, taking full advantageof the rich correlation of the source in the MD encoding anddecoding, and achieving satisfactory performance at all ratesand redundancies.

One of the motivations of investigating such a scheme isShannon’s comments on reliable communications [15, pp. 75]:“Any redundancy in the source will usually help if it is utilizedat the receiving point. In particular, If the source already hasa certain redundancy a sizable fraction of the letters canbe received incorrectly and still reconstructed by the context.”Since MDC can be understood as a joint source channel codingproblem, one would expect similar effect exists that the inherentdependence in the source can be utilized.

Fig. 1. Forward and inverse time-domain lapped transforms (TDLT).

The potential benefit of utilizing the source correlation inpractical MD encoding and decoding suggests a design that de-viates from the traditional transform paradigm, for which theerror resilient design of the lapped transform [16] is a suit-able platform. In [17] and [18], the lapped orthogonal transformoutput is split at the block level for transmission. The transformis designed to introduce some correlations to help estimatinglost blocks at the decoder. In [19], a new family of transformscalled the time-domain lapped transform (TDLT) [20] is used,which simplifies the design. The Wiener filter is applied in [21]to further improve the performance. However, these methodssacrifice too much coding efficiency to get good error resilience,and they all suffer from the prediction residual at high rates. Inaddition, they have to change the transform to adjust the redun-dancy, which is not convenient in practice.

In this paper, we present a new MDC scheme by combiningthe time-domain lapped transform, block level source splitting,Wiener filter based prediction, and the encoding of the predic-tion residual. We formulate the joint optimization of all com-ponents in the system, analyze its asymptotic performance, andgive various design results. Our scheme resolves the problemsof previous methods, and outperforms the MMDSQ, RD-MDC,and PCT in MD image coding, especially at low redundancies.In addition to the performance gain, the proposed method alsohas lower complexity than many existing methods such as thePCT, GPCT, and RD-MDC.

II. PROBLEM FORMULATION AND OPTIMAL DESIGN

In this section, after a brief introduction of the time-domainlapped transform (TDLT) developed in [20], we present theproposed MDC scheme and discuss its advantages over othermethods. We then formulate the joint design of various compo-nents of the system and derive the optimal solution.

A. Time-Domain Lapped Transform

Fig. 1 shows the block diagram of the TDLT developed in[20], which is a low-cost extension of the DCT, but with compet-itive performance compared to JPEG 2000 [22]. At the encoder,an prefilter is employed at the boundary of two blocks( is the block size). The -point DCT is then applied toeach block, creating basis functions that cover two blocks. In thedecoder, the inverse DCT and postfilter at block boundaries

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 3: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

SUN et al.: MULTIPLE DESCRIPTION CODING WITH PREDICTION COMPENSATION 1039

Fig. 2. Block diagrams of encoding (a) and decoding (b) one description in the proposed method.

are applied. Matrices and have the following structures toyield near-optimal linear-phase lapped transform [20]:

(1)

(2)

(3)

where and are identity matrix and counteri-dentity matrix, respectively. Matrix is an in-vertible matrix that can be optimized for different purposes. Inthis paper, denotes a block diagonal matrix withmatrices and on the diagonal and zeros elsewhere.

Denote and as the first and the last rows of theprefilter , and and as the first and the last columnsof , respectively, i.e.,

(4)

(5)

the forward lapped transform and inversetransform can be written as

(6)

(7)

B. Overview of the Proposed MDC Scheme and Its Advantages

In this paper, we modify the TDLT framework to generatetwo descriptions. Fig. 2 illustrates the encoding and decodingof one description by the proposed method. The other descrip-tion is obtained in a symmetric manner. In this paper, we use

, , and to denote the th block of prefilter input,DCT input, and DCT output. The reconstruction of a variable isdenoted by the hat operator.

In Fig. 2, the prefiltered blocks are split into even-in-dexed blocks and odd-indexed blocks. After DCT, quantization,and entropy coding, the two groups form the base layers ofthe two descriptions, respectively. Since the base layer only in-cludes half of the input, each description also contains an en-hancement layer to help reconstructing the other half, if one de-scription is lost. To fully exploit the source correlation, we use

a Wiener filter to predict a missing block using its two recon-structed neighboring blocks, as in [21]. More precisely, inFig. 2 is predicted using the nearest samples fromand samples from , where . The deriva-tion of the Wiener filter is given in the Appendix. Different from[21], in this paper, the prediction residuals are furtherDCT-transformed, quantized and entropy-coded to form the en-hancement layer of each description.

At the decoder side, if one description is lost, the missingblocks are first estimated from the received base layer blocks byWiener filter. The decoded enhancement layer blocks are thenadded to the estimation before postfiltering. When both descrip-tions are available, only the decoded base layers from the twodescriptions are fed to the postfilter for reconstruction. The en-hancement layer in each description is simply discarded. There-fore, the enhancement layer is the redundancy introduced by ourmethod, which can be easily controlled by adjusting the quanti-zation step size.

Our scheme enjoys various advantages over existing methods.Compared to the MDSQ/MMDSQ, RD-MDC, PCT/GPCT, ituses Wiener filter to exploit the source correlation betweenneighboring blocks in MD encoding and decoding. The benefitof utilizing the source correlation can be seen in Section V,where even a simple system with DCT and linear interpolationbased prediction can outperform other methods in some cases,especially at low redundancies. Compared with the MDSQand [8]–[10], our method offers a more flexible control of theredundancy. Our method is also superior to [11]–[13] by usingpredictive coding, as it is well established that for correlatedsources, predictive coding has better rate-distortion perfor-mance than direct encoding [23, pp. 113], i.e., it achieves lowerdistortion at the same bit rate.

The proposed scheme also avoids other limitations of the PCTand GPCT. First, the block level splitting has much better codingefficiency than coefficient level splitting, as the structure withineach block is intact. Second, prediction in the spatial domain al-lows the residual to be coded even at low redundancies, whereimportant image edge information can still survive the exces-sive quantization after the prediction and the DCT. In contrast,most coefficients in the PCT/GPCT are split directly in thiscase, which are uncorrelated and cannot be predicted from each

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 4: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

1040 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009

others. Third, as shown in Section III, at high redundancies, thetheoretical performance of our method is as good as the GPCT.In practical image coding, our method outperforms not only thePCT/GPCT, but also the MMDSQ and RD-MDC, as our methodis well suited for nonstationary signals. Finally, our scheme canbe easily optimized and implemented. Only one prefilter andone Wiener filter are needed, and the prediction residuals areprocessed in the same way as base layer blocks.

C. Joint Optimal Design of the System Components

Let [in bits/pixel (bpp)] represent the overall bit rate ofthe two descriptions, i.e., the ratio between the total bits of thetwo descriptions and the number of input samples. Let and

denote the average bits for each base layer sample and eachenhancement layer sample, respectively. The bit rate for eachdescription is thus bpp/description, and the totalrate is .

As shown below, given the target bit rate , the probabilityof losing one description, the matrix in prefilter , and

the size of the Wiener filter, we can find the closed-form ex-pressions of the corresponding optimal Wiener filter and bitallocation and that minimize the expected distortion atthe receiver. To further find the optimal matrix with the min-imal expected distortion, an unconstrained numerical optimiza-tion program, such as the function fminunc in MATLAB, can beused, by treating all entries of as unknown variables. Next,we give the derivation of the first step, i.e., the optimal and

for a given . The derivation of the corresponding Wienerfilter is given in the Appendix.

Let and be the central distortion and side distortion,i.e., the mean squared error (MSE) when the decoder receivestwo and one description, respectively. The expected distortion

is defined as

(8)

In the proposed MDC scheme, each description contains halfbase layer blocks and half enhancement layer blocks. Only baselayer blocks are used if two descriptions are received. If one ofthem is lost, half of the input is reconstructed via base layer andthe other half is via enhancement layer. Therefore

(9)

where and are the MSE caused by the subbandquantization noise in base layer and prediction-compensatedenhancement layer, respectively. The subscript denotes theblock size (we use to represent the DPCM case, asstudied in Section III). Substituting into (8), we have

(10)

We first find the optimal expressions of andfor given and , under the optimal bit allocation withineach block. Since the inverse lapped transform is generallynot orthogonal, the reconstruction error or is the

weighted combination of subband quantization noises, and theweighting parameters are the norms of the inverse transformfilters. Let the quantization noise of be . After theinverse TDLT, the reconstruction error becomes . Asusual, we assume the quantization noises of different subbandsare uncorrelated. The MSE of the reconstruction is thus

(11)

where is the variance of the th entry of , and isthe th column of . Under the assumptions of high rates andi.i.d. sources, can be written as [23, pp. 108]

(12)

where is a constant that depends on the input statistics and thequantization scheme. is the bits allocated to the th entryof a base layer block, and . isthe variance of the th entry of , which is the th diagonalelement of the autocorrelation matrix

(13)

and .Upon optimal bit allocation of [24], the minimal value

for (11) is given by

(14)This is, in fact, the objective function of the single descriptioncoding. For block transforms, the minimum value of (14) isachieved by the Karhunen–Loève transform (KLT). For lappedtransforms and longer filter banks, there is no closed-form solu-tion, but numerical optimization method can be used to find thesolution that minimizes (14).

In single description coding, a performance measure calledthe coding gain is defined based on

(15)

When , the coding gain of the DCT and the optimizedTDLT in [20] is 8.83 and 9.62 dB, respectively, for first-orderGaussian–Markov inputs with correlation .

We now look at , the MSE caused by enhancementlayer blocks in the side decoder. It can be seen from Fig. 2 thatthe prediction residual is

(16)

where is the Wiener filter-based prediction of fromand . The derivation of for the given

prefilter is given in the Appendix. At the decoder, the recon-struction of a predictively coded block is

(17)

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 5: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

SUN et al.: MULTIPLE DESCRIPTION CODING WITH PREDICTION COMPENSATION 1041

where is the reconstruction of . From (16) and (17),we have the following:

(18)

In other words, the reconstruction error of is equal to thatof the prediction residual . This is, indeed, a property ofany differential coding system [23, pp. 113].

As in Fig. 2, let be the DCT transform of , we have

(19)

where is the quantization noise of . After post-filtering, the reconstruction error becomes

, and its MSE is

(20)

where is the bits allocated to the th entry of . isthe variance of the th entry of , given by the th diagonalelement of the autocorrelation matrix . The expression of

is given by (44) in the Appendix.Since (20) has the same format as (11), the derivation from

(11) to (14) can be applied here, and the minimal value ofafter optimal bit allocation is, therefore

(21)The remaining bit allocation issue is to find the optimal and

for the given that minimize the expected distortion in(10). Substituting (14) and (21) into (10), the problem can bewritten as

(22)

This can be solved using the Lagrangian method, and the op-timal bit allocation can be found to be

(23)

At high rates, i.e., if is not forced to 0 in (23), substitutingthis into (14) and (21), we have

(24)

In this case, the minimal objective function in (22) becomes

(25)

Finally, plugging (24) into and in (9) yields the followingdistortion product :

(26)

which is further discussed in the DPCM case in Section III.The following remarks are in order.Remark 1: Equation (23) shows that more bits should

be allocated to the prediction residual when the loss prob-ability is higher or when is larger (the data aremore difficult to predict). Notice that when

. In this case, the methodreduces to our previous approach in [21]. However, it should beemphasized that this threshold is derived based on the first-orderGauss–Markov model. For nonstationary signals like naturalimages, we show in Section V that sending prediction residualis beneficial even at very low bit rates and low redundancies,because these bits are spent at regions with strong edges, andcan thus significantly improve the reconstruction quality.

Remark 2: Equation (25) shows that the optimal TDLT for theproposed MDC scheme needs to minimize , whereasthe optimal single description transform should minimizein (14). Since and are dependent, the optimal TDLTfor the proposed MDC is different from the single descriptioncase, although the difference is not much, as shown in Sec-tion IV.

Remark 3: Equation (25) also shows that when the loss prob-ability changes, we always need to minimize .Therefore, the optimal transform is independent of . This is de-sired in practice. If is forced to 0, (25) would be invalid, andthe optimal transform would be a function of . However, ouroptimization results in Section IV show that the optimal trans-form is not sensitive to .

Remark 4: The DCT can be viewed as a special case of theTDLT when the prefilter is . In this case, the derivationabove is still valid. The performance of our method in the DCTcase is also studied in Sections IV and V.

III. ASYMPTOTIC PERFORMANCE OF THE PROPOSED METHOD

We now analyze the asymptotic performance of our trans-form-based method. This is achieved by studying the DPCMcase, which has the same performance as transform coding whenthe block size goes to infinity [24]. In the DPCM case, we splitthe data into even and odd samples, and use DPCM to encodeeach group. Each description also predicts the samples in theother group and encodes the residual as the enhancement layer.Since the block size is , no block transform or lappedtransform is used.

This special case of our method is indeed an improvementof Jayant’s DPCM based method in [8] and [9]. As discussedin Section I, Jayant’s method splits the source at sample leveland uses linear prediction between the two descriptions, but itdoes not encode the prediction residual, which is handled by theenhancement layer in our method.

A. Optimization in the DPCM Case

Assume the input follows a first-order Gauss–Markov modelwith correlation coefficient . After splitting, each part is still

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 6: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

1042 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009

a first-order Gauss–Markov signal, but with correlation coeffi-cient . Let and be two consecutive samples in onedescription. If the DPCM is used in each description, the optimalprediction of from is given by . Athigh rates, the variance of the residual is

(27)

where is the quantization noise variance of and is neg-ligible at high rates.

Similar to (18), the DPCM system also satisfies[23]. Therefore, the reconstruction error

of after DPCM with bit rate is

(28)

This can be viewed as the counterpart for (14).When one description is lost, each missing sample is first

predicted from and , i.e.,

(29)

where the optimal solution for is the Wiener filter. At highrates, it is approximately [9]

(30)

In this case, the variance of the residual is

(31)

In Jayant’s method, this error exists even at high rates. In ourmethod, the prediction residual is further encoded at rate inthe enhancement layer in each description, leading to a reduceddistortion of

(32)

which is the counterpart for (21). Since (28) and (32) have thesame format as (14) and (21), respectively, the bit allocationsolution (23) to (22) can be applied here as well. From (27) and(31), we know that , so (23) becomes

(33)

B. Asymptotic Performance of the Proposed Method

It is shown in [3] (see also [6]) that under the high rate as-sumption, the product of the side distortion and the central dis-tortion of an MDC scheme for a stationary source satisfies

(34)

where is the entropy power of the source [23, pp. 95]. Thisproperty of the distortion product has been widely used as aperformance measure for MDC.

In the DPCM case of our method, substituting the bit alloca-tion into (28), (32) and (9), we can get

(35)This corresponds to (26) for . To gain more insight,notice that the variance of the innovation sequence of a first-order Gauss–Markov signal is , which is alsothe entropy power of the signal, i.e., . Therefore, thedistortion product in (35) can be written as

(36)

Comparing (36) and (34), the performance of the proposedmethod is away from the theoretical bound by a factor of

, or about 3 dB for small values of . This is similarto the performance of the GPCT in [14] at high redundancies.However, it should be noted that this result is for the first-orderGauss–Markov signal, which is stationary. We show in Sec-tion V that for nonstationary signals such as natural images,the proposed method performs equally well as other MDCalgorithms at high redundancies. For low redundancies, it sig-nificantly outperforms other methods, because the block-levelprediction compensation is quite suitable for nonstationarysignals.

Fig. 3 compares the expected distortion of our method andJayant’s method for . and are also included. Theinput is a first-order Gauss–Markov source with . Theconstant is chosen as 1. The figure shows that the side SNR ofJayant’s method could not be improved at high rates, due to theexistence of prediction error (31), which dominates the expecteddistortion at high rates. This problem is resolved in our methodby encoding the prediction error. Thus, both the side SNR andthe expected SNR can be improved as the increase of the bitrate. However, this is achieved at the price of reduced centralSNR. Fig. 3 also shows that when the bit rate is below about 2.6bits/sample, no residual is encoded, and our method reduces toJayant’s method. However, as shown in Section V, for naturalimages, it is beneficial to encode the residuals even at very lowbit rates.

IV. OPTIMIZATION RESULTS AND DESIGN EXAMPLES

In this section, we show various optimized distortion prod-ucts and filters for the proposed method. Table I summarizes theoptimized distortion product under different configura-tions of the transform, block size , description loss probability

, and Wiener filter size. The input is chosen as a first-orderGauss–Markov signal with and . As in Fig. 3,

is selected. The bit rate is bpp. The source codes togenerate the results in this paper can be downloaded from [25].

Three configurations of the proposed algorithm are comparedwith the DPCM in Table I. The first one jointly optimizes theprefilter and the Wiener filter to minimize the expected dis-tortion in (10). We denote this method as multiple descriptionlapped transform with prediction compensation (MDLT-PC).The second one is denoted as TDLT-PC, which uses the bestprefilter for single description coding, i.e., by optimizing (14).The last one is denoted as DCT-PC, which only uses the DCT,

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 7: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

SUN et al.: MULTIPLE DESCRIPTION CODING WITH PREDICTION COMPENSATION 1043

Fig. 3. Comparison of the expected distortion � of the DPCM case of ourmethod and Jayant’s method. The central and side distortions are also includedfor reference (� � ���, � � �, � � ����, and � � �).

TABLE IDISTORTION PRODUCT � � ���� � OF DIFFERENT CONFIGURATIONS

(WITH � � �, � � ����, � � �, AND � � � bpp)

i.e., no prefilter. In TDLT-PC and DCT-PC, Wiener filter andresidual encoding are still used.

In Table I, MDLT-PC already approaches the performance ofDPCM when . For , the distortion product ofthe three configurations is about 5%, 10%, and 20% inferior tothe DPCM, respectively. Reducing Wiener filter size increasesthe distortion, but the change is less than 3%. For , evenMDLT-PC with has better performance than TDLT-PCwith , showing the advantage of joint optimization. Notealso that MDLT-PC with is worse than DCT-PC with

by about 2.5%. Finally, as shown in [21], Wiener filterwith is optimal for the DCT. In this case, it actuallyreduces to linear interpolation. Despite its simplicity, we showin Section V that the DCT-PC can still achieve better resultsthan the MMDSQ and RD-MDC in some cases.

We show in Section II that the optimal transform is indepen-dent of when . Simulation results also show that thecoding gain of the optimized transform does not change muchwhen is forced to 0. The optimal result is also not sensitive to

and . For , when the rate and the error probabilityvary in a large range, the coding gain of the optimized TDLT

only changes between 9.41 and 9.61 dB. This makes it possible

to fix the transform and still achieves near optimal performancefor all practical scenarios.

Two design examples are used in the image coding in the nextsection. The first one is optimized for , bpp,

and . The coding gain of the result is 9.53 dB,the product is 0.00164, and the corresponding optimizedmatrix in the TDLT prefilter is given by

(37)

The second example is optimized for , bpp,and , with a coding gain of 9.54 dB and. The optimized matrix is

(38)

In this case, the corresponding 8 2 Wiener filter is simply

V. PERFORMANCE IN MD IMAGE CODING

In this section, we evaluate the performance of the proposedmethod in the MD coding of natural images. The source codesand testing parameters for all results can be downloaded from[25]. Six 512 512 standard test images with different charac-teristics are used. The block size is chosen to be 8. Two de-scriptions are generated by partitioning the transformed blocksin a checkerboard pattern. All base layer blocks in each descrip-tion are grouped together to form a 256 512 subimage, whichis then encoded by the embedded entropy coding in [22]. Simi-larly, all enhancement layer blocks in each description are alsogrouped together and encoded by the same entropy coding al-gorithm.

The 1-D Wiener filter is applied to 2-D images in a separablemanner. First, each row of a block is Wiener-predicted usingthe co-located rows in the left and right neighboring blocks.Second, each column of the block is estimated using theco-located columns in the top and bottom neighboring blocks.After that, the average of the row and the column predictions isused as the final prediction of the 2-D block. The 2-D predictionresidual is then calculated and encoded in the enhancementlayer.

We first study the tradeoff between and , the cen-tral PSNR and the side PSNR. This is related to the distor-tion product and has been used as a performance mea-sure in, e.g., [3], [5], [6], and [13]. In our method, this is easilyachieved by varying the quantization step sizes of the two layers.In Figs. 4 and 5, our method is compared with the wavelet andTarp filter-based MMDSQ in [6], and the JPEG 2000-basedRD-MDC in [13] (source code at [26]), which represent the stateof the art in MDC. The total bit rate is and bpp,respectively. Four configurations of the proposed method aretested, namely, MDLT-PC with and (37), (38),

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 8: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

1044 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009

Fig. 4. Performances of MDLT-PC, MMDSQ and RD-MDC at � � � bit/pixel. (a) Barbara. (b) Boat. (c) Baboon. (d) Goldhill. (e) Lena. (f) Peppers.

Fig. 5. Performances of MDLT-PC, MMDSQ and RD-MDC at � � ���� bit/pixel. (a) Barbara. (b) Boat. (c) Baboon. (d) Goldhill. (e) Lena. (f) Peppers.

DLT-PC with and DCT-PC with . The average ofthe two side MSEs is used to calculate the side PSNR in thesefigures.

It can be seen that the performances of the MMDSQ andRD-MDC are quite similar in many cases. The MMDSQ per-forms better for smooth images like Lena and Peppers, whereas

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 9: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

SUN et al.: MULTIPLE DESCRIPTION CODING WITH PREDICTION COMPENSATION 1045

Fig. 6. Side decoder results with total rate of 1 bpp in (a)–(c) and 0.25 bpp in (d)–(f). The central PSNR isincluded in the parentheses. (a) Barbara by MMDSQ:24.52 dB (36.09 dB). (b) Barbara by RD-MDC: 27.60 dB (36.07 dB). (c) Barbara by MDLT-PC: 31.68 dB (36.07 dB). (d) Goldhill by MMDSQ: 24.58 dB(30.35 dB). (e) Goldhill by RD-MDC: 24.46 dB (30.35 dB). (f) Goldhill by MDLT-PC: 27.22 dB (30.35 dB).

the RD-MDC is better for images with more textures, such asBarbara.

At the same central PSNR, the side PSNR of our method out-performs the MDSQ and RD-MDC in most cases of Barbara,Boat, Baboon, and Goldhill, many with large margins, espe-cially at low redundancies. The improvement can be more than8 and 6 dB for bpp and bpp, respectively. Forsmooth images like Lena and Peppers, our method still achievesbetter performance in many low redundancy cases. At high re-dundancies, the three methods behave similarly, especially forsmooth images.

The results also show that the proposed method is not sen-sitive to the size of the Wiener filter. In most cases, the low-cost MDLT-PC with has similar performance to theMDLT-PC with , making it a good candidate for prac-tical applications, due to its simpler Wiener filter.

When the single description optimized TDLT is used (TDLT-PC), the curves are lower than the jointly optimized MDLT-PCcurves by less than 0.5 dB in all cases. If only the DCT is used(DCT-PC), the curves can be up to 2 dB lower, with more degra-dations at low rates. These relationships agree with Table I forGauss–Markov sources. Note that even the simple DCT-PC withlinear interpolation can achieve similar or better performancethan the MMDSQ and RD-MDC at some low redundancy ex-periments, demonstrating the benefit of utilizing source corre-lation in these cases.

The first points of all curves of our method correspond to. In this case, our method reduces to the prediction-only

method in [21], and reasonable side PSNR is still obtained. Incontrast, the side decoding performance of the MMDSQ withlow redundancies is not satisfactory, because most received bitsare in the second layer and cannot be used. Similar defect alsoexists in the RD-MDC at low redundancies, because there islittle information about half of the code-blocks.

Fig. 6 shows some decoding results with one description.The three methods are compared at the same total bit rate andsame central PSNR. Clearly, our method achieves significantimprovement in both the side PSNR and the visual quality.

Finally, we compare our method and the PCT. In Fig. 9 of[7], the of image Lena is kept at 35.78 dB. At this ,the JPEG-based single description coder in [7] needs a rate of0.60 bpp, whereas the TDLT codec in [7] only needs 0.346 bpp,due to the improved transform and entropy coding. The PCTachieves a side PSNR of 27.94 and 29.63 dB, with redundancyof 15% and 22% over 0.60 bpp, respectively. To get the same

, our MDLT-PC with only needs a redundancy of9.88% and 16.02% over 0.346 bpp, respectively. In addition,the lowest PCT redundancy in Fig. 9 of [7] is about 10%, with

dB. When , our method can achieve a redun-dancy of 4.38%, and the corresponding is 25.45 dB. There-fore, our method can achieve the same with less redundancythan the PCT, and it also reaches a lower redundancy range. Thecomparison is not very rigorous because of the codec differ-ences, but even if the PCT can be implemented in the TDLTframework, it is still difficult to achieve similar performance toour method due to its aforementioned limitations.

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 10: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

1046 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009

VI. CONCLUSION

This paper presents an MDC paradigm by integrating time-domain lapped transform, block level splitting, linear predictionand compensation. The method provides effective redundancycontrol while simultaneously takes advantage of the source cor-relation. Image coding results show that it outperforms existingmethods significantly.

The findings in this paper can be useful for other applica-tions such as MD video coding. Moreover, the proposed methodcan be further improved. For example, the enhancement layerbit rate can be reduced by refining the entropy coding for theresidual, and 2-D prediction and adaptive filter can be appliedto reduce the prediction residual, as in [27].

The method can also be generalized to create descriptions,where . A direct generalization is to split the image into

subimages after prefiltering. Each subimage is coded as thebase layer of one description. Each description also encodes theprediction residuals of all other subimages as the enhancementlayer. Further improvement of this approach is our ongoing re-search.

APPENDIX

In this section, we derive the Wiener filter in Fig. 2. In [21],all data in the two neighboring blocks are used to estimate a lostblock. Since the performance is not sensitive to the size of theWiener filter, in this paper we use neighboringsamples from each side to estimate the lost block, where canbe chosen to obtain a desired tradeoff between complexity andperformance.

In Fig. 2, the prediction of is , whereis the prediction filter, and is a vector con-taining nearest neighboring samples next to , sam-ples from and samples from . That is

(39)

where

(40)

The autocorrelation of the prediction residual is

(41)

The Wiener filter that minimizes the MSE is, where is the correlation be-

tween and , and is the autocorrelation .At high rate, the quantization noise can be ignored [21], [23, pp.114], and the Wiener filter can be approximated as

(42)

The matrices involved can be obtained from the structure of thelapped transform in Fig. 1. Define

(43)

As shown in Fig. 1, . Thus,, where , and

is the autocorrelation matrix of . In this paper,is obtained by assuming the input follows a first-order

Gauss–Markov model with correlation coefficient .Matrices and in (42) can then be obtainedfrom the appropriate sub-matrices of .

From and the Wienerfilter given in (42), we have

(44)

In this paper, this is used in (20) to obtain the distortion of theenhancement layer.

As in [21], we normalize the Wiener filter to have unit rowsums. In addition, two Wiener filters are used at theboundary to predict a block from only the top or the bottomneighboring block. Their derivations are similar to (42).

ACKNOWLEDGMENT

The authors would like to thank the anonymous reviewers fortheir suggestions that have significantly enhanced the presenta-tion of this paper.

REFERENCES

[1] V. K. Goyal, “Multiple description coding: Compression meets the net-work,” IEEE Signal Process. Mag., vol. 18, no. 5, pp. 74–93, Sep. 2001.

[2] V. A. Vaishampayan, “Design of multiple description scalar quan-tizers,” IEEE Trans. Inf. Theory, vol. 39, no. 3, pp. 821–834, May1993.

[3] V. A. Vaishampayan and J.-C. Batllo, “Asymptotic analysis of multipledescription quantizers,” IEEE Trans. Inf. Theory, vol. 44, no. 1, pp.278–284, Jan. 1998.

[4] J.-C. Batllo and V. A. Vaishampayan, “Asymptotic performance ofmultiple description transform codes,” IEEE Trans. Inf. Theory, vol.43, no. 2, pp. 703–707, Mar. 1997.

[5] S. D. Servetto, K. Ramchandran, V. A. Vaishampayan, and K.Nahrstedt, “Multiple description wavelet based image coding,” IEEETrans. Image Process., vol. 9, no. 5, pp. 813–826, May 2000.

[6] C. Tian and S. S. Hemami, “A new class of multiple description scalarquantizer and its application to image coding,” IEEE Signal Process.Lett., vol. 12, no. 4, pp. 329–332, Apr. 2005.

[7] Y. Wang, M. T. Orchard, V. A. Vaishampayan, and A. R. Reibman,“Multiple description coding using pairwise correlating transforms,”IEEE Trans. Image Process., vol. 10, no. 3, pp. 351–366, Mar. 2001.

[8] N. S. Jayant, “Subsampling of a DPCM speech channel to providetwo self-contained half-rate channels,” Bell Syst. Tech. J., vol. 60, pp.501–509, Apr. 1981.

[9] N. S. Jayant and S. W. Christensen, “Effects of packet losses in wave-form coded speech and improvements due to an odd-even sample-inter-polation procedure,” IEEE Trans. Commun., vol. COM-29, no. 2, pp.101–109, Feb. 1981.

[10] A. Ingle and V. A. Vaishampayan, “DPCM system design for diversitysystems with applications to packetized speech,” IEEE Trans. SpeechAudio Process., vol. 3, no. 1, pp. 48–58, Jan. 1995.

[11] W. Jiang and A. Ortega, “Multiple description coding via polyphasetransform and selective quantization,” in Proc. SPIE Conf. Visual Com-munication Image Processing, Feb. 1999, vol. 3653, pp. 998–1008.

[12] A. C. Miguel, A. E. Mohr, and E. A. Riskin, “SPIHT for generalizedmultiple description coding,” in Proc. IEEE Conf. Image Processing,Oct. 1999, vol. 3, pp. 842–846.

[13] T. Tillo, M. Grangetto, and G. Olmo, “Multiple description imagecoding based on lagrangian rate allocation,” IEEE Trans. ImageProcess., vol. 16, no. 3, pp. 673–683, Mar. 2007.

[14] Y. Wang, A. R. Reibman, M. T. Orchard, and H. Jafarkhani, “Animprovement to multiple description transform coding,” IEEE Trans.Signal Process., vol. 50, no. 11, pp. 2843–2854, Nov. 2002.

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.

Page 11: IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY 2009 1037 Multiple ...jiel/papers/MDLTPC.pdf · 2009-04-07 · IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 5, MAY

SUN et al.: MULTIPLE DESCRIPTION CODING WITH PREDICTION COMPENSATION 1047

[15] C. E. Shannon and W. Weaver, The Mathematical Theory of Commu-nication. Champaign, IL: Univ. Illinois Press, 1998.

[16] H. S. Malvar, Signal Processing With Lapped Transforms. Norwood,MA: Artech House, 1992.

[17] S. S. Hemami, “Reconstruction-optimized lapped transforms for robustimage transmission,” IEEE Trans. Circuits Syst. Video Technol., vol. 6,no. 2, pp. 168–181, Apr. 1996.

[18] D. Chung and Y. Wang, “Lapped orthogonal transform designedfor error resilient image coding,” IEEE Trans. Circuits Syst. VideoTechnol., vol. 12, no. 9, pp. 752–764, Sep. 2002.

[19] C. Tu, T. D. Tran, and J. Liang, “Error resilient pre-/post-filtering forDCT-based block coding systems,” IEEE Trans. Image Process., vol.15, no. 1, pp. 30–39, Jan. 2006.

[20] T. D. Tran, J. Liang, and C. Tu, “Lapped transform via time-domainpre- and post-processing,” IEEE Trans. Signal Process., vol. 51, no. 6,pp. 1557–1571, Jun. 2003.

[21] J. Liang, C. Tu, L. Gan, T. D. Tran, and K.-K. Ma, “Wiener filter-basederror resilient time domain lapped transform,” IEEE Trans. ImageProcess., vol. 16, no. 2, pp. 428–441, Feb. 2007.

[22] C. Tu and T. D. Tran, “Context based entropy coding of block transformcoefficients for image compression,” IEEE Trans. Image Process., vol.11, no. 11, pp. 1271–1283, Nov. 2002.

[23] D. S. Taubman and M. W. Marcellin, JPEG 2000: Image CompressionFundamentals, Standards, and Practice. Boston, MA: Kluwer, 2002.

[24] N. S. Jayant and P. Noll, Digital Coding of Waveforms: Principles andApplications to Speech and Video. Englewood Cliffs, NJ: Prentice-Hall, 1984.

[25] MDLT-PC Source Code [Online]. Available: http://www.ensc.sfu.ca/~jiel/MDLTPC.html

[26] RD-MDC Source Code [Online]. Available: http://www.telem-atica.polito.it/sas-ipl

[27] J. Liang, X. Li, G. Sun, and T. D. Tran, “Two-dimensional Wienerfilters for error resilient time domain lapped transform,” in Proc. IEEEInt. Conf. Acoustics, Speech, Signal Processing, Toulouse, France, May2006, vol. III, pp. 241–244.

Guoqian Sun received the B.S. and M.S. degreesfrom Southeast University, China, in 2000 and 2003,respectively. He is currently pursuing the M.A.Sc.degree at the School of Engineering Science, SimonFraser University, Burnaby, BC, Canada.

He is currently with the Service Provider VideoTechnology Group, Cisco Systems, Inc., Vancouver,BC. From 2003 to 2005, he was with ZTE Corpora-tion, Nanjing, China. His research interests includemultimedia communications, image/video compres-sions, and digital communications.

Upul Samarawickrama (S’05) received the B.Sc. degree in electronics andtelecommunications engineering from the University of Moratuwa, Sri Lanka,in 2002, and the M.Sc. degree from the University of Manitoba, Canada, in 2005.He is currently pursuing the Ph.D. degree at the School of Engineering Science,Simon Fraser University, Burnaby, BC, Canada.

His research interests include multiple description coding, joint source-channel coding, and digital communications.

Jie Liang (S’99–M’04) received the B.E. and M.E.degrees from Xi’an Jiaotong University, China,in 1992 and 1995, the M.E. degree from NationalUniversity of Singapore (NUS) in 1998, and thePh.D. degree from The Johns Hopkins University,Baltimore, MD, in 2003, respectively.

Since May 2004, he has been an Assistant Pro-fessor at the School of Engineering Science, SimonFraser University, Burnaby, BC, Canada. From 2003to 2004, he was with the Video Codec Group of Mi-crosoft Digital Media Division, Redmond,WA. His

research interests include image and video coding, multirate signal processing,and joint source channel-coding.

Chao Tian (S’00–M’05) received the B.S. degreein electronic engineering from Tsinghua University,Beijing, China, in 2000, and the M.S. and Ph.D.degrees in electrical and computer engineering fromCornell University, Ithaca, NY, in 2003 and 2005,respectively.

He was a postdoctoral researcher at Ecole Poly-technique Federale de Lausanne (EPFL) from 2005to 2007. He joined AT&T Labs-Research, FlorhamPark, NJ, in 2007, where he is now a Senior Memberof Technical Staff. His research interests include mul-

tiuser information theory, joint source-channel coding, quantization design andanalysis, as well as image/video coding and processing.

Chengjie Tu (S’02–M’04) received the B.E. and M.E. degrees from Universityof Science and Technology of China in 1994 and 1997, respectively, and theM.S.E. and Ph.D. degrees from The Johns Hopkins University, Baltimore, MD,in 2002 and 2003, respectively.

Since 2004, he has been with the Video Codec Group, Microsoft Corporation,Redmond, WA. His current research interests include image/video compression,rich media presentation/communication, and multirate signal processing.

Trac D. Tran (S’94–M’98–SM’08) received the B.S.and M.S. degrees from the Massachusetts Institute ofTechnology, Cambridge, in 1993 and 1994, respec-tively, and the Ph.D. degree from the University ofWisconsin, Madison, in 1998, all in electrical engi-neering.

In July 1998, he joined the Department of Elec-trical and Computer Engineering, The Johns HopkinsUniversity, Baltimore, MD, where he currently holdsthe rank of Associate Professor. His research inter-ests are in the field of digital signal processing, par-

ticularly in compressed sensing, sampling, multirate systems, filter banks, trans-forms, wavelets, and their applications in signal analysis, compression, pro-cessing, and communications.

Prof. Tran was the Co-Director (with Prof. J. L. Prince) of the 33rd An-nual Conference on Information Sciences and Systems (CISS’99), Baltimore,MD, in March 1999. In the summer of 2002, he was an ASEE/ONR SummerFaculty Research Fellow at the Naval Air Warfare Center-Weapons Division(NAWCWD), China Lake, CA. He has served as Associate Editor of the IEEETRANSACTIONS ON SIGNAL PROCESSING as well as the IEEE TRANSACTIONS

ON IMAGE PROCESSING. He currently serves as a member of the IEEE TechnicalCommittee on Signal Processing Theory and Methods (SPTM TC). He receivedthe NSF CAREER award in 2001 and the William H. Huggins Excellence inTeaching Award from The Johns Hopkins University in 2007.

Authorized licensed use limited to: IEEE Xplore. Downloaded on April 6, 2009 at 21:56 from IEEE Xplore. Restrictions apply.