1248 ieee transactions on image processing, vol. 19, no. … motion aligned auto-regressive model...

11
1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010 A Motion-Aligned Auto-Regressive Model for Frame Rate Up Conversion Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang, and Wen Gao, Fellow, IEEE Abstract—In this paper, a motion-aligned auto-regressive (MAAR) model is proposed for frame rate up conversion, where each pixel is interpolated as the average of the results generated by one forward MAAR (Fw-MAAR) model and one backward MAAR (Bw-MAAR) model. In the Fw-MAAR model, each pixel in the to-be-interpolated frame is generated as a linear weighted summation of the pixels within a motion-aligned square neighbor- hood in the previous frame. To derive more accurate interpolation weights, the aligned actual pixels in the following frame are also estimated as a linear weighted summation of the newly interpolated pixels in the to-be-interpolated frame by the same weights. Consequently, the backward-aligned actual pixels in the following frame can be estimated as a weighted summation of the corresponding pixels within an enlarged square neighborhood in the previous frame. The Bw-MAAR is performed likewise except that it is operated in the reverse direction. A damping Newton algorithm is then proposed to compute the adaptive interpolation weights for the Fw-MAAR and Bw-MAAR models. Extensive experiments demonstrate that the proposed MAAR model is able to achieve superior performance than the traditional frame interpolation methods such as MCI, OBMC, and AOBMC, and it is even better than STAR model for the most test sequences with moderate or large motions. Index Terms—Damping Newton algorithm, frame rate up con- version (FRUC), motion-aligned auto-regressive (MAAR) model, motion compensation interpolation (MCI). I. INTRODUCTION F RAME RATE up conversion (FRUC) is a statistical inter- polation procedure where pixels in the to-be-interpolated frame are generated based on observations from the previous and following frames. FRUC is commonly used in format con- version, e.g., from 24 fps film content to 30 fps video content, or from 50 fields to 60 fields per second. FRUC is also used to enhance the visual quality of reconstructed frames in low bit Manuscript received October 10, 2008; revised November 24, 2009. First published December 22, 2009; current version published April 16, 2010. This work was supported in part by the National Science Foundations of China (60736043 and 60803068), in part by the Major State Basic Research Devel- opment Program of China (973 Program 2009CB320904), and in part by the France Telecom Project (OR_PED-BBR). The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Jenq-Neng Hwang. Y. Zhang and D. Zhao are with the Department of Computer Science, Harbin Institute of Technology, Harbin 150001, China (e-mail: [email protected]; [email protected]). S. Ma and W. Gao are with the School of Electronics Engineering and Computer Science, Peking University, Beijing 100871, China (e-mail: [email protected]; [email protected]). R. Wang is with France Telecom R&D, Beijing Co., Ltd., China (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TIP.2009.2039055 rate video coding, where only some key frames in the original sequence are encoded and all the remaining frames are interpo- lated by the adjacent decoded key frames. In addition, FRUC is of other uses such as video surveillance, medical imaging and remote sensing. One of the simplest FRUC techniques is the combination of adjacent video frames, like frame repetition or frame averaging. In static regions it performs quite well; but in moving regions, when the motion between successive frames is neglected, the interpolated frame becomes very choppy. A more effective and prevalent technique is motion com- pensation interpolation (MCI), which performs interpolation along the motion trajectory between the previous and following frames. Since motion information is utilized in MCI, the ac- curacy of motion vector determines the quality of interpolated frames. To improve the accuracy of motion vector between successive frames, many pioneering works have been done, which can be divided into two categories: motion estimation (ME) and motion vector post processing. In [1], de Haan et al. proposed a 3-D recursive search (3-DRS) method to get the true motions between successive frames. Choi et al. [2] proposed a bi-directional ME method to obtain more reliable motions for FRUC. A hierarchical ME [3] was also proposed by Lee et al. to obtain more faithful motions. To overcome the limitation of constant velocity assumption, a constant acceleration model [4] was adopted by Gan et al. to further improve the accuracy of motion vector between successive frames. Since the actual pixels are absent in the to-be-interpolated frame, the ME in FRUC is performed between the previous frame and the fol- lowing frame, which may sometimes result in nonconsistent motion fields. Consequently, some motion vector post pro- cessing methods [5], [6] were proposed to smooth the motion fields. These MCI methods mentioned above, usually performed block by block, have achieved much better performance than frame averaging and frame repetition. Whereas block edges may not always be consistent with the heterogeneous object edges, and thus blocking artifacts are usually perceived in the regions where one block has a significantly different motion compared with its neighbors. By extending traditional MCI, overlapped block motion com- pensation (OBMC) [7] was employed in FRUC for its efficiency of reducing blocking artifacts. For example, Zhai et al. [8] pro- posed an OBMC method to suppress the blocking artifacts by positioning overlapped blocks from the previous and following frames. Compared with MCI, OBMC is able to generate a much smoother interpolated frame. However, OBMC may result in blurring or over-smoothing artifacts in case of nonconsistent motion regions since it assigns fixed weights for neighboring 1057-7149/$26.00 © 2010 IEEE Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Upload: others

Post on 13-May-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

A Motion-Aligned Auto-Regressive Modelfor Frame Rate Up Conversion

Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang, and Wen Gao, Fellow, IEEE

Abstract—In this paper, a motion-aligned auto-regressive(MAAR) model is proposed for frame rate up conversion, whereeach pixel is interpolated as the average of the results generatedby one forward MAAR (Fw-MAAR) model and one backwardMAAR (Bw-MAAR) model. In the Fw-MAAR model, each pixelin the to-be-interpolated frame is generated as a linear weightedsummation of the pixels within a motion-aligned square neighbor-hood in the previous frame. To derive more accurate interpolationweights, the aligned actual pixels in the following frame arealso estimated as a linear weighted summation of the newlyinterpolated pixels in the to-be-interpolated frame by the sameweights. Consequently, the backward-aligned actual pixels in thefollowing frame can be estimated as a weighted summation of thecorresponding pixels within an enlarged square neighborhood inthe previous frame. The Bw-MAAR is performed likewise exceptthat it is operated in the reverse direction. A damping Newtonalgorithm is then proposed to compute the adaptive interpolationweights for the Fw-MAAR and Bw-MAAR models. Extensiveexperiments demonstrate that the proposed MAAR model isable to achieve superior performance than the traditional frameinterpolation methods such as MCI, OBMC, and AOBMC, and itis even better than STAR model for the most test sequences withmoderate or large motions.

Index Terms—Damping Newton algorithm, frame rate up con-version (FRUC), motion-aligned auto-regressive (MAAR) model,motion compensation interpolation (MCI).

I. INTRODUCTION

F RAME RATE up conversion (FRUC) is a statistical inter-polation procedure where pixels in the to-be-interpolated

frame are generated based on observations from the previousand following frames. FRUC is commonly used in format con-version, e.g., from 24 fps film content to 30 fps video content,or from 50 fields to 60 fields per second. FRUC is also usedto enhance the visual quality of reconstructed frames in low bit

Manuscript received October 10, 2008; revised November 24, 2009. Firstpublished December 22, 2009; current version published April 16, 2010. Thiswork was supported in part by the National Science Foundations of China(60736043 and 60803068), in part by the Major State Basic Research Devel-opment Program of China (973 Program 2009CB320904), and in part by theFrance Telecom Project (OR_PED-BBR). The associate editor coordinating thereview of this manuscript and approving it for publication was Dr. Jenq-NengHwang.

Y. Zhang and D. Zhao are with the Department of Computer Science, HarbinInstitute of Technology, Harbin 150001, China (e-mail: [email protected];[email protected]).

S. Ma and W. Gao are with the School of Electronics Engineering andComputer Science, Peking University, Beijing 100871, China (e-mail:[email protected]; [email protected]).

R. Wang is with France Telecom R&D, Beijing Co., Ltd., China (e-mail:[email protected]).

Color versions of one or more of the figures in this paper are available onlineat http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TIP.2009.2039055

rate video coding, where only some key frames in the originalsequence are encoded and all the remaining frames are interpo-lated by the adjacent decoded key frames. In addition, FRUC isof other uses such as video surveillance, medical imaging andremote sensing.

One of the simplest FRUC techniques is the combination ofadjacent video frames, like frame repetition or frame averaging.In static regions it performs quite well; but in moving regions,when the motion between successive frames is neglected, theinterpolated frame becomes very choppy.

A more effective and prevalent technique is motion com-pensation interpolation (MCI), which performs interpolationalong the motion trajectory between the previous and followingframes. Since motion information is utilized in MCI, the ac-curacy of motion vector determines the quality of interpolatedframes. To improve the accuracy of motion vector betweensuccessive frames, many pioneering works have been done,which can be divided into two categories: motion estimation(ME) and motion vector post processing. In [1], de Haan et al.proposed a 3-D recursive search (3-DRS) method to get the truemotions between successive frames. Choi et al. [2] proposed abi-directional ME method to obtain more reliable motions forFRUC. A hierarchical ME [3] was also proposed by Lee et al.to obtain more faithful motions. To overcome the limitation ofconstant velocity assumption, a constant acceleration model[4] was adopted by Gan et al. to further improve the accuracyof motion vector between successive frames. Since the actualpixels are absent in the to-be-interpolated frame, the ME inFRUC is performed between the previous frame and the fol-lowing frame, which may sometimes result in nonconsistentmotion fields. Consequently, some motion vector post pro-cessing methods [5], [6] were proposed to smooth the motionfields.

These MCI methods mentioned above, usually performedblock by block, have achieved much better performance thanframe averaging and frame repetition. Whereas block edgesmay not always be consistent with the heterogeneous objectedges, and thus blocking artifacts are usually perceived in theregions where one block has a significantly different motioncompared with its neighbors.

By extending traditional MCI, overlapped block motion com-pensation (OBMC) [7] was employed in FRUC for its efficiencyof reducing blocking artifacts. For example, Zhai et al. [8] pro-posed an OBMC method to suppress the blocking artifacts bypositioning overlapped blocks from the previous and followingframes. Compared with MCI, OBMC is able to generate a muchsmoother interpolated frame. However, OBMC may result inblurring or over-smoothing artifacts in case of nonconsistentmotion regions since it assigns fixed weights for neighboring

1057-7149/$26.00 © 2010 IEEE

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 2: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

ZHANG et al.: MOTION-ALIGNED AUTO-REGRESSIVE MODEL FOR FRAME RATE UP CONVERSION 1249

blocks. For better adjusting the weights of OBMC, an adaptiveOBMC (AOBMC) [9] was also proposed to tune the weights ofdifferent blocks according to the reliability of neighboring mo-tion vectors.

All of the aforementioned works are dedicated to find themost similar blocks in the previous and following frames, andthen take the combination of the similar blocks as the ultimateinterpolation results. In most cases, the motion displacementresolution is fractional-pel rather than integer-pel. As a result,most of these methods first interpolate the previous and fol-lowing frames to half-pel or quarter-pel resolution with a fixedinterpolation filter and then search the most similar blocks.However, as stated in [10], the invariant interpolation filterdoes not consider the nonstationary statistical properties ofvideo signals, which will result in aliasing or displacementestimation errors in the interpolation process. Thus, all theseFRUC methods will inevitably suffer from the inferiority of thefixed interpolation filter.

Considering the inherent shortcomings of the fixed inter-polation filter, it would be beneficial to devise an adaptiveinterpolation scheme in FRUC. Auto-regressive (AR) model[11], which is an efficient, compact description of randomprocess, has attracted a lot of attention in image and videoapplications thanks to its desirable abilities of adaptively ad-justing the weighting parameters to local pixel structure. ARpredictors exploiting spatial information [11] and temporalinformation [12], [13] have been proposed to improve theprediction performance. In [11], the predictive pixels wereestimated from the past observations by a piecewise 2-D ARmodel. In [12], AR model was utilized to interpolate a highresolution sequence based on the low resolution one. In [13],AR was employed to predict one pixel as a linear combinationof pixels in a spatio-temporal neighborhood. AR was alsoemployed to detect and interpolate the missing data in imagesbased on its property of handling fine detail and accounting forintensity variation [14], [15]. Furthermore, AR was exploitedto perform ME [16] and super-resolution [17].

In our earlier work [18], we used a spatio-temporal auto-re-gressive (STAR) model to perform FRUC, where each pixelin the interpolated frame is approximated as the weightedcombination of a sample space including the pixels within itstwo temporal neighborhoods from the previous and followingframes as well as the available interpolated pixels within itsspatial neighborhood in the current to-be-interpolated frame.In the STAR model, the temporal neighborhoods are cen-tered at the collocated pixels in the previous and followingframes. Consequently, for the sequences with moderate or largemotions, the collocated regions among the previous frame,the to-be-interpolated frame and the following frame will bemuch different, which will cause performance degradation andmeanwhile bring blurring and ghost artifacts in the interpolatedframes. Although the neighborhood with large size can coverthe corresponding most similar pixels in the previous andfollowing frames, it will also contain many uncorrelated pixels,which will decrease the performance and greatly increase thecomputational complexity.

In this paper, a motion-aligned AR (MAAR) model is pro-posed to improve the performance of FRUC by interpolating one

pixel as the weighted combination of pixels within the temporalneighborhoods, which are centered at the motion-aligned pixelsin the previous and following frames. First, for each to-be-in-terpolated block, the corresponding motion-aligned blocks inthe previous and following frames are found by integer-pel ac-curacy ME. Then, the interpolation process is carried out witha forward MAAR (Fw-MAAR) interpolation and a backwardMAAR (Bw-MAAR) interpolation. In Fw-MAAR interpola-tion, each pixel in the to-be-interpolated block is generated asa linear weighted summation of pixels within a motion-alignedsquare neighborhood in the previous frame. Based on the as-sumption that the sample covariance does not change in the mo-tion-aligned blocks, the Fw-MAAR interpolation is propagatedalong the temporal axis. In other words, the newly generatedpixels in the to-be-interpolated block are further used to esti-mate the aligned actual pixels in the following frame by the sameweights. Consequently, each actual pixel within the backwardmotion-aligned block in the following frame can be estimatedas a weighted summation of pixels within an enlarged squareneighborhood in the previous frame. The Bw-MAAR interpola-tion performs likewise except that it is operated in the reverse di-rection. A damping Newton algorithm is then devised to obtainthe interpolation weights for the Fw-MAAR and Bw-MAARmodels. The ultimate pixels in the to-be-interpolated block arethen generated as the average of the Fw-MAAR and Bw-MAARinterpolation results.

The main differences between the contributions of this paperand [18] are as follows: 1) in this paper, we perform the regres-sion along the aligned motions rather than regressing along theco-located position as in [18]; 2) in this paper, only one forward(Fw-MAAR) or one backward (Bw-MAAR) temporal neigh-borhood is utilized to perform interpolation, and then the av-erage of Fw-MAAR and Bw-MAAR is generated as the ultimateinterpolation. Whereas in [18], besides the forward and back-ward temporal neighborhoods, one spatial neighborhood is alsoutilized to perform interpolation; 3) only three frames (includingthe current to-be-interpolated frame, one previous frame andone following frame) are involved in Fw-MAAR or Bw-MAAR,whereas five frames (including the current to-be-interpolatedframe, one previous frame, one following frame, the next to-be-interpolated frame and the original frame which follows the nextto-be-interpolated frame) are necessary for STAR in [18]; 4) theoptimization objective in this paper is to minimize the differencebetween the previous (or following) frame and its approximatedframe via interpolation propagation. However, in [18], the opti-mization objective is to minimize the summation of three differ-ences: a) the interpolation difference of the current to-be-inter-polated frame between two iterations; b) the interpolation dif-ference of the next to-be-interpolated frame between two itera-tions; c) the difference between the original frame, which is lo-cated between the current to-be-interpolated frame and the nextto-be-interpolated frame, and its approximated frame.

The remainder of this paper is organized as follows. Section IIdescribes the proposed MAAR model for FRUC. Section IIIpresents the proposed damping Newton algorithm for weight es-timation in detail. Experimental results and analysis conductedon various sequences are given in Section IV. Finally, a briefconclusion is given in Section V.

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 3: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1250 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

Fig. 1. Proposed MAAR model diagram for FRUC.

Fig. 2. Motion aligning.

II. MOTION-ALIGNED AUTO-REGRESSIVE MODEL

The goal of the proposed MAAR model is to interpolate oneintermediate frame with high visual quality. In this section, theMAAR model diagram is presented first, and the model descrip-tion is given afterward.

A. MAAR Model Diagram

The proposed MAAR model diagram for FRUC is illus-trated in Fig. 1. First, the to-be-interpolated frame is dividedinto nonoverlapped blocks and then the motion aligning isperformed. The motion aligning process is depicted in Fig. 2,where for each block in the to-be-interpolated frame ,we find its aligned blocks in the previous frame and thefollowing frame by integer-pel ME. Various MEmethods can be adopted to find the aligned blocks. In thispaper, we utilize bi-directional ME [2] to obtain the forwardand backward motion vectors. After motion aligning, the for-ward and backward damping Newton algorithms are performedto derive the corresponding weights for the Fw-MAAR andBw-MAAR models, respectively. Finally, the pixels within thecurrent to-be-interpolated block are generated by averaging theinterpolation results obtained by the newly derived Fw-MAARand Bw-MAAR weights.

B. Model Description

Let be a rectangular block within the to-be-interpolated frame , be the forward motion-alignedblock in the previous frame and be the backwardmotion-aligned block in the following frame .can be interpolated by the proposed MAAR model based on thepixels in (by Fw-MAAR) as well as the pixels in(by Bw-MAAR).

Fig. 3. Fw-MAAR model (with supporting order � � �).

Fig. 4. Bw-MAAR model (with supporting order � � �).

The Fw-MAAR model is depicted in Fig. 3, where each pixelin is interpolated as a linear weighted summation of thepixels within a square neighborhood centered at the forward mo-tion-aligned pixel in . The radius (represented by ) of thesquare neighborhood is defined to be the supporting order of theFw-MAAR model. Thus, when the supporting order is set to be1, the neighborhood of the Fw-MAAR model is a 3 3 square,which is shown as the gray circles and the solid cross in Fig. 3.Due to the piecewise stationary characteristics of natural im-ages, the Fw-MAAR weights corresponding to each pixel withinone block are assumed to be the same. Similarly, the Bw-MAARmodel is depicted in Fig. 4, where each pixel in the to-be-inter-polated block is interpolated as a linear weighted summation ofthe pixels within the backward motion-aligned square neighbor-hood in the following frame.

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 4: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

ZHANG et al.: MOTION-ALIGNED AUTO-REGRESSIVE MODEL FOR FRAME RATE UP CONVERSION 1251

Fig. 5. Propagation of Fw-MAAR interpolation along the temporal axis.

Since the Fw-MAAR and Bw-MAAR models are quitesimilar, we will take the Fw-MAAR model as an example inthe following description. Denote by the interpo-lated pixel located at within . According to theFw-MAAR model, can be interpolated as

(1)

where represents the corresponding pixel within theforward motion-aligned block , is the supporting orderof the Fw-MAAR model, is the Fw-MAAR weight and

is the additive white Gaussian noise.Equation (1) can also be expressed in a matrix-vector form as

(2)

whererepresents the intensity values in the to-be-inter-

polated block,represents the intensity values in the forward mo-

tion-aligned block, is theFw-MAAR weight vector and is a function transferring

to a matrix. Here, the th row of ma-trix consists of neighboring pixels of the thpixel within , . The Fw-MAAR weight

should be chosen the “best” in some sense. In this paper, themean square error (MSE) is used to measure its performance

(3)

where denotes the norm. The optimal Fw-MAARweights can be computed by minimizing the MSE in (3).However, since the actual pixels in are not availablein FRUC, (3) cannot be directly used to compute the weights.To solve this problem, an iterative algorithm using a dampingNewton method is proposed in the following section.

III. DAMPING NEWTON METHOD FOR WEIGHT ESTIMATION

A. Propagation of Fw-MAAR Interpolation

Based on the Fw-MAAR model described in Section II, wecontinue the Fw-MAAR interpolation along the temporal axis.That is to say, assuming that we have obtained the pixels inthe to-be-interpolated frame by the Fw-MAAR model,we apply the same Fw-MAAR model to further estimate theactual pixels within the backward motion-aligned block in thefollowing frame. This process is illustrated in Fig. 5.

It is a reasonable assumption that the Fw-MAAR weights,used to estimate the backward motion-aligned actual pixelswithin frame , remain the same as those used to inter-polate the pixels within frame since there is a highsimilarity between pixels along the motion trajectory fromframe to . As shown in the right part of Fig. 5, theactual pixels in the backward motion-aligned block withinframe can be estimated as

(4)

where is a column vector representing the concate-nated and lexicographically ordered intensity values in the back-ward motion-aligned block within frame . isa matrix, whose element is computed ac-cording to (2), and is the same as that in (2). Incorporating(2) into (4), we can get the interpolation of utilizing thecorresponding pixels within the forward motion-aligned block

as

(5)

where is a function transferring the column vectorto a matrix, represents the new weightvector corresponding to the enlarged square neighborhood ofsized shown as the gray circles and thesolid cross in Fig. 6. Here, the th row of matrix consistsof neighboring pixels of the th pixel within ,

.In Fig. 6, each pixel within the backward motion-aligned

block in the following frame can be estimated as a

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 5: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1252 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

Fig. 6. MAAR interpolation between the previous and following frames afterthe propagation of MAAR interpolation in the intermediate frame.

weighted summation of the pixels within an enlarged squareneighborhood in the previous frame . It should be noted thatthe length of the new weight vector corresponding to theenlarged square is and each element of is thequadratic of the elements of . In other words, can beexpressed in a 2-D form as

(6)

with and .

B. Damping Newton Algorithm

After we get (5), the weight vector can be computed byminimizing the following MSE:

(7)

where is defined to be theresidual vector. Assume to be the approximation of , weexpress around in Tayer series as

(8)

where is the Jacobian matrix of at . Here, the Jaco-bian matrix can be computed as

...

.... . .

... (9)

with

(10)

where represents the th row of the matrix with.

TABLE ISUMMARY OF THE PROPOSED DAMPING NEWTON ALGORITHM

Let ; we can easily compute as

(11)

According to the Newton iteration algorithm, if we have ob-tained the weight vector after the th iteration, the weightvector in the th iteration can be updated as

(12)

Since the direct Newton iteration method imposes a strong de-mand on the initial value of , we modified (12) by adding adamping factor as

(13)

where with isthe damping factor. By adding the damping factor, the conver-gence properties of the proposed algorithm can be improved.

The proposed damping Newton algorithm to estimate theweight vector is summarized in Table I. As shown in thesummary of the iterative algorithm, the forward MSEis used to judge whether the damping Newton algorithm hasbeen converged or not. In other words, if is smallerthan a preset threshold , it is considered that the dampingNewton algorithm has been converged and the iteration canbe preterminated, otherwise, the algorithm moves to the nextiteration. In the experiment, the threshold is empiricallyset to be 5.

If we define the change of the weight vector comparedto as

(16)then the convergence ratio of each weight is defined as

(17)

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 6: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

ZHANG et al.: MOTION-ALIGNED AUTO-REGRESSIVE MODEL FOR FRAME RATE UP CONVERSION 1253

Fig. 7. Average PSNRs of 50 frames interpolated by the proposed MAAR model with different initial weights. (a) Mobile (QCIF). (b) Tempete (CIF).

TABLE IISUMMARY OF THE DAMPING FACTOR ADJUSTING

Based on the convergence ratios in (17), we propose a methodto adaptively adjust the damping factor, which is summarized inTable II.

In Table II, parameters , and can prevent the weightstrapping into the sub-optimal solutions by adjusting the updateof the weights in each iteration. It is noted that , and shouldbe positive and should satisfy , and . Inthis paper, the values of , and are empirically set to be 0.7,0.2, and 2, respectively, and is set to be the identity matrix

. The computation of is synchronous with the computationof , and consequently in Table II is the same with thatin Table I.

and should be given proper values, since they playa significant role on the convergence of the proposed dampingNewton algorithm. In this paper we set to be the inter-polation results by the traditional FRUC methods, such as MCI,OBMC or AOBMC. After that we interpolate the correspondingpixels within the motion-aligned blocks in frames and

by the Fw-MAAR model as

(18)

(19)

The initial Fw-MAAR weight vector is computed accordingto

(20)

After deriving the Fw-MAAR weight , the similar iterativeprocess can also be applied to obtain the Bw-MAAR weight .Finally, the interpolated pixels within the current to-be-interpo-lated block are computed as

(21)

IV. EXPERIMENTAL RESULTS AND ANALYSIS

In this section, various experiments are conducted to showthe performance of the proposed MAAR model. In the first sub-section, we will study the convergence property of the proposedMAAR model. In the second subsection, the interpolation per-formance comparisons are presented.

A. Convergence Study

In this subsection, Mobile (QCIF) and Tempete (CIF) se-quences are selected to conduct the experiments to show theconvergence property of the proposed MAAR model. Therespective block sizes for Mobile (QCIF) and Tempete (CIF)are set to be 8 8 and 16 16. The supporting orders are set tobe 2 for these two test sequences. Every other frame of the first100 frames in each test sequence is dropped and interpolatedby the proposed MAAR model.

The interpolation results of MCI, OBMC [8], and AOBMC[9] are selected to compute the initial weights of the proposedMAAR model. It is noted that the block size of MCI, OBMCand AOBMC is set to be 8 8, and the block size of MAAR isvarying according to the sequence content. The motion vectorsof MCI, OBMC and AOBMC are the same and they are de-rived by the bi-directional ME method in [2]. The 8 8 blockof MAAR is motion aligned directly by the motion vectors ofMCI, OBMC and AOBMC, and for the block larger than 88, it is motion aligned by averaging the motion vectors of all

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 7: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1254 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

TABLE IIIPSNRs OF THE INTERPOLATED FRAMES BY DIFFERENT METHODS

the 8 8 blocks within it. The average PSNRs of 50 interpo-lated frames in Mobile (QCIF) and Tempete (CIF), generated bythe proposed MAAR model with different initial weights, areplotted in Fig. 7. It is observed that with the increase of iterationnumber, the PSNRs of the interpolated frames increase greatlyat first and then tend to be stable. Another observation is thatthe convergence of the proposed damping Newton algorithm isindependent of the initial weights, because the PSNRs tend tobe converged with the increase of iteration number regardless ofthe initial weights.

The experimental results shown in Fig. 7 provide the em-pirical evidence on the convergence of the proposed dampingNewton algorithm. Based on these observations, the maximumiteration number is set to be 3 for all the experiments in the fol-lowing subsection.

B. Interpolation Comparison

In this subsection, various simulations are performed in fiveQCIF sequences, five CIF sequences, five 4CIF sequences andfive 720P sequences to show the performance of the proposedMAAR model. The 15-Hz frames are interpolated into 30 Hzfor all the test sequences (except for 720P sequences, which areinterpolated to 60 from 30 Hz). As in the previous subsection,every other frame of the first 100 frames in each test sequence isdropped and then interpolated by the proposed MAAR model,MCI, OBMC [8], AOBMC [9], and STAR [18], respectively. In[18], the search range of MCI, OBMC, and AOBMC is set tobe 4 for QCIF and CIF sequences and it is set to be 8 for 4CIFand 720P sequences. While in this paper, the search range is setto be 4 for all the test sequences. The average PSNRs of the 50interpolated frames within each test sequence are presented inTable III.

In Table III, Fw represents the interpolation results generatedby Fw-MAAR model, Bw by Bw-MAAR model, Averageby averaging Fw and Bw. represents the block size

and supporting order of the MAAR model. Thehighest PSNR values, among all the methods, of each testsequence are represented in bold type. It can be observed thatthe performance made only by Fw-MAAR or by Bw-MAARis not as good as the best result among MCI, OBMC, AOBMCand STAR. However, when Fw-MAAR and Bw-MAAR are av-eraged, the performance improves significantly over all the testsequences. The average gains achieved by the proposed MAARmodel (initialized with AOBMC) are 0.17, 0.53, and 0.47 dBcompared with the best results among MCI, OBMC, AOBMC,and STAR for the CIF, 4CIF, and 720P sequences, respectively.For QCIF sequences, MAAR has about 0.01 dB loss comparedwith STAR; however, it still has about 0.62 dB gain comparedwith the best one among MCI, OBMC, and AOBMC. OBMCand AOBMC have better performance than MCI due to theutilizing of weighted motion compensation. STAR model isvery efficient for some highly textured sequences [e.g., Mo-bile (QCIF) and Flower (CIF)]. This is because these highlytextured sequences are full of nonrigid objects, which are hardto be recovered by traditional block based MCI, OBMC , andAOBMC method. Compared with MAAR, STAR model hasone additional spatial neighborhood, and consequently it isvery suitable for those highly textured sequences with finedetails. Besides, STAR also achieves good results for sequenceswith low motion [e.g., Foreman (QCIF) and Mother (QCIF)].However, for the sequences with moderate and large motions,the performance of STAR drops quickly. For example, there isabout 0.96 dB loss for Foreman (CIF) compared with MAAR.This is because the temporal neighborhoods of the STARmodel are centered at the collocated pixels in the previous and

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 8: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

ZHANG et al.: MOTION-ALIGNED AUTO-REGRESSIVE MODEL FOR FRAME RATE UP CONVERSION 1255

Fig. 8. PSNR comparisons over each interpolated frame with different initial weights for Mobile (QCIF) and Spincalendar (720p). (a) MCI method and theproposed MAAR model initialized with MCI. (b) OBMC method and the proposed MAAR model initialized with OBMC. (c) AOBMC method and the proposedMAAR model initialized with AOBMC.

following frames, and it can achieve good results for static orlow motion sequence. However, with the increase of motion thedifference between the collocated regions between successiveframes (especially the first and the fifth frame) becomes larger,which will lead to the drop of performance. On the contrary,the temporal neighborhoods of the MAAR model are motionaligned, which can ensure the pixels within the temporal neigh-borhoods are very similar to the target pixel. Consequently,the performance of MAAR model is still very high for thesequences with moderate or large motions.

Several findings have been derived from the experiments.First, when is larger than 2, the performance can not beimproved, whereas the computational complexity is greatlyincreased. Second, larger achieves better performance over720P sequences and other sequences with complex texture e.g.,Mobile (QCIF), Tempete (QCIF), Tempete (CIF), and Flower(CIF). Thus, is set to be 2 for these sequences and it is setto be 1 for the remaining sequences. Third, for the sequenceswith larger motion, small results in better performance, e.g.,

is 8 for Foreman (CIF) and Crossstreet (4CIF) due to the

fact that smaller blocks can better describe more active regionswith larger motions. For 720P sequences full of homogeneousregions with little or no motion, large will achieve betterperformance, e.g., is 64 for Bigships and Spincalendar. Forother sequences, moderate block sizes (between 8 8 and 64

64) achieve sufficiently good performance.Fig. 8 presents the PSNRs of each interpolated frame in Mo-

bile (QCIF) and Spincalendar (720p). It can be easily observedthat no matter what interpolation method is used to computethe initial weights, the proposed MAAR model achieves higherPSNR than the corresponding interpolation method over eachinterpolated frame. In particular, the PSNR gain is up to 2.5 dBover the frames around the 25th frame in Mobile (QCIF) and upto 3.0 dB over the frames around the 15th frame in Spincalendar(720p).

The local area zoom-in comparisons over Spincalendar(720P) and News (CIF) are shown in Figs. 9 and 10. In Fig. 9,the number “11” in the two rows can not be seen clearly in theinterpolated frames generated by MCI, OBMC and AOBMC.In the STAR result, the number “11” in the bottom row can be

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 9: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1256 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

Fig. 9. Zoom-in comparison of the 27th interpolated frame in Spincalendar (720P). (a) Original. (b) STAR result. (c) The proposed MAAR model initializedwith MCI. (d) MCI result. (e) The proposed MAAR model initialized with OBMC. (f) OBMC result. (g) The proposed MAAR model initialized with AOBMC.(h) AOBMC result.

Fig. 10. Zoom-in comparison of the fifth interpolated frame in News (CIF). (a) Original. (b) STAR result. (c) The proposed MAAR model initialized with MCI.(d) MCI result. (e) The proposed MAAR model initialized with OBMC. (f) OBMC result. (g) The proposed MAAR model initialized with AOBMC. (h) AOBMCresult.

observed clearly; however, there are still some ghost artifactsaround number “11” in the top row. This is because number“11” is spinning, and consequently the collocated regionsamong the interpolated frame, the previous frame and thefollowing frame are relatively different, which results in someweak ghost artifacts around number “11”. Whereas, in theinterpolated frames by the proposed MAAR model the number“11” in both rows can be clearly perceived, e.g., Fig. 9(c), (e),and (g). In Fig. 10, significant ghost artifacts can be observedaround the left leg of the actress in the interpolated frames byMCI, OBMC [8], and AOBMC [9]. On the contrary, the ghostartifacts are greatly reduced in the interpolated frames by theproposed MAAR model, no matter what method is utilized tocompute the initial weights, e.g., Fig. 10(c), (e), and (g). Inthe interpolated frame by STAR, the ghost artifact is tamperedto some extent but still significant, this is because the left legof the actress is rotating, which can not be recovered well bySTAR model. However, the eyes and nose area of the actressis relatively clear, which shows the efficiency of STAR forthose highly textured details. It is obvious that the proposedMAAR model achieves much better visual quality than MCI,OBMC and AOBMC do. Compared with STAR, MAAR canalso achieve better visual quality on the areas with moderate

or large motion although it may have some loss on highlytextured details. This is greatly attributed to the property thatthe proposed MAAR model is able to tune the interpolationweights according to the local region property along the alignedmotion.

Table IV gives the average processing times (seconds/frame)on a typical computer (2.5–GHz Intel Dual Core, 2 G Memory).Since MCI, OBMC and AOBMC employ the same ME methodin the experiment, they require more or less the same processingtime. The computational complexities of STAR and MAAR aremuch higher than those of MCI, OBMC, and AOBMC due tothe inverse matrix operation involved in these two methods. Thecomputational complexity of STAR depends on the levels ofmotion activity, and, thus, the average processing time of largemotion sequence is much higher than that of static or smallmotion sequence. The processing time of the proposed MAARmodel is about 4 and 19 times longer than that of MCI when

is set to be 1 and 2, respectively. However, the proposedMAAR model is still faster than STAR model. Although thecomputational complexity is higher, the proposed MAAR modelachieves much better quality than traditional methods (MCI,OBMC, and AOBMC) for all the interpolated frames. To speedup the proposed damping Newton algorithm, we can look for

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 10: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

ZHANG et al.: MOTION-ALIGNED AUTO-REGRESSIVE MODEL FOR FRAME RATE UP CONVERSION 1257

TABLE IVCOMPARISON OF AVERAGE PROCESSING TIMES (s/frame)

optimization algorithms such as the fast algorithm proposed byX. Wu [11] which can greatly reduce the computational com-plexity. In addition, since (8) is convex, it has a global minimumwhich can be found using the steepest descent method as in [17].Further studies on reducing the computational complexity of theproposed MAAR model are needed.

V. CONCLUSION

A motion-aligned auto-regressive (MAAR) model is pro-posed to perform frame interpolation for frame rate upconversion in this paper. The integer-pel accuracy motionvector of each to-be-interpolated block is first obtained byperforming traditional motion estimation between the previousand following frames. Along the aligned motion, the inter-polation of each block is carried out by one forward MAAR(Fw-MAAR) and one backward MAAR (Bw-MAAR) models.In the Fw-MAAR model, each pixel is interpolated as a linearweighted summation of pixels within a square neighborhood,centered at the forward motion-aligned pixel in the previousframe. The Fw-MAAR interpolation is then propagated alongthe aligned motion and consequently each backward mo-tion-aligned pixel in the following frame can be estimated asa weighted summation of pixels within an enlarged squareneighborhood in the previous frame. A damping Newton al-gorithm is then designed to compute the interpolation weightsof the Fw-MAAR model. The interpolation and correspondingweight derivation of the Bw-MAAR model are similar to thoseof the Fw-MAAR model. The ultimate interpolated pixel iscomputed as the average of the interpolation results generatedby the Fw-MAAR and Bw-MAAR models.

Experimental results show that the proposed MAAR modelis able to adaptively tune its interpolation weights accordingto the local properties of image signals along the aligned mo-tion. Compared with the traditional interpolation methods, such

as MCI, OBMC, and AOBMC, the proposed MAAR modelimproves the visual quality of the interpolated frames whilemaintaining higher PSNR values. Although it has small losswhen compared with STAR for some highly textured or lowmotion sequences, it is better for the most test sequences withmoderate or large motions. One shortcoming of the proposedMAAR model is the computation complexity required to cal-culate the interpolation weights; however, the increased com-plexity does not prevent its use for off-line processing. In thenext step, we plan to investigate faster algorithms to simplifythe MAAR weights estimation.

REFERENCES

[1] G. de Haan, P. W. Biezen, H. Huijgen, and O. A. Ojo, “True-motion es-timation with 3-D recursive search block matching,” IEEE Trans. Cir-cuits Syst. Video Technol., vol. 3, no. 5, pp. 368–379, Oct. 1993.

[2] B. T. Choi, S. H. Lee, and S. J. Ko, “New frame rate up-conversionusing bi-directional motion estimation,” IEEE Trans. Consum. Elec-tron., vol. 46, no. 3, pp. 603–609, Aug. 2000.

[3] G. I. Lee, B. W. Jeon, R. H. Park, and S. H. Lee, “Hierarchical motioncompensated frame rate up-conversion based on the Gaussian/Lapla-cian pyramid,” in Proc. IEEE Int. Conf. Consumer Electronics, 2003,pp. 350–351.

[4] Z. Gan, L. Qi, and X. Zhu, “Motion compensated frame interpolationbased on H.264 decoder,” Electron. Lett., vol. 43, no. 1, pp. 96–98, Jan.2007.

[5] R. Castagno, P. Haavisto, and G. Ramponi, “A method for motionadaptive frame rate up-conversion,” IEEE Trans. Circuits Syst. VideoTechnol., vol. 6, no. 5, pp. 436–446, Oct. 1996.

[6] G. Dane and T. Nguyen, “Motion vector processing for frame rate upconversion,” in Proc. IEEE Int. Conf. Acoustics, Speech, and SignalProcessing, 2004, pp. 309–312.

[7] B. Girod, “Efficiency analysis of multihypothesis motion-compensat-edprediction for video coding,” IEEE Trans. Image Process., vol. 9, no.2, pp. 173–183, Feb. 2000.

[8] J. Zhai, K. Yu, J. Li, and S. Li, “A low complexity motion compensatedframe interpolation method,” in Proc. ISCAS, May 2005, vol. 5, pp.4927–4930.

[9] B. D. Choi, J. W. Han, C. S. Kim, and S. J. Ko, “Motion-compensatedframe interpolation using bilateral motion estimation and adaptiveoverlapped block motion compensation,” IEEE Trans. Circuits Syst.Video Technol., vol. 17, no. 4, pp. 407–416, Apr. 2007.

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.

Page 11: 1248 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. … Motion Aligned Auto-regressive Model f… · Yongbing Zhang, Debin Zhao, Siwei Ma, Student Member, IEEE, Ronggang Wang,

1258 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 5, MAY 2010

[10] T. Wedi, “Adaptive interpolation filters and high-resolution displace-ments for video coding,” IEEE Trans. Circuits Syst. Video Technol.,vol. 16, no. 4, pp. 484–491, Apr. 2006.

[11] X. Wu, K. U. Barthel, and W. Zhang, “Piecewise 2D autoregression forpredictive image coding,” in Proc. IEEE Int. Conf. Image Processing,1998, pp. 901–904.

[12] Y. Zheng and X. Li, “Video modeling via spatio-temporal adaptive lo-calized learning (STALL),” in Proc. Signals, Systems and Computers,,2006, pp. 779–783.

[13] X. Li, “Least-square prediction for backward adaptive video coding,”EURASIP J. Appl. Signal Process., vol. 2006, pp. 1–13, 2006.

[14] A. C. Kokaram, R. D. Morris, W. J. Fitzgerald, and P. J. W. Rayner,“Detection of missing data in image sequences,” IEEE Trans. ImageProcess., vol. 4, no. 11, pp. 1496–1508, Nov. 1995.

[15] A. C. Kokaram, R. D. Morris, W. J. Fitzgerald, and P. J. W. Rayner,“Interpolation of missing data in image sequences,” IEEE Trans. ImageProcess., vol. 4, no. 11, pp. 1509–1519, Nov. 1995.

[16] S. N. Efstratiadis and A. K. Katsaggelos, “A model-based pel-recur-sive motion estimation algorithm,” in Proc. IEEE Int. Conf. Acoustics,Speech, and Signal Processing, 1990, pp. 1973–1976.

[17] X. Zhang and X. Wu, “Image interpolation by adaptive 2-D autore-gressive modeling and soft-decision estimation,” IEEE Trans. ImageProcess., vol. 17, no. 6, pp. 887–896, Jun. 2008.

[18] Y. Zhang, D. Zhao, X. Ji, R. Wang, and W. Gao, “A spatio-temporalauto-regressive model for frame rate up conversion,” IEEE Trans. Cir-cuits Syst. Video Technol., vol. 19, no. 9, pp. 1289–1301, Sep. 2009.

Yongbing Zhang received the B.A. degree in Eng-lish and the M.S. degree in computer science fromthe Harbin Institute of Technology, Harbin, China,in 2004 and 2006, respectively. He is currentlypursuing the Ph.D. degree at the Harbin Institute ofTechnology.

His research interests include video processing,image and video coding, video streaming, andtransmission.

Debin Zhao received the B.S., M.S., and Ph.D. de-grees in computer science from the Harbin Institute ofTechnology, Harbin, China, in 1985, 1988, and 1998,respectively.

He is now a Professor in the Department of Com-puter Science, Harbin Institute of Technology. He haspublished over 200 technical articles in refereed jour-nals and conference proceedings in the areas of imageand video coding, video processing, video streamingand transmission, and pattern recognition.

Dr. Zhao was the recipient of three National Sci-ence and Technology Progress Awards of China (Second Prize).

Siwei Ma (S’03) received the B.S. degree from Shan-dong Normal University, Jinan, China, in 1999, andthe Ph.D. degree in computer science from the Insti-tute of Computing Technology, Chinese Academy ofSciences, Beijing, China, in 2005.

From 2005 to 2007, he was a postdoctorate withthe University of Southern California, Los Angeles.Then he joined the Institute of Digital Media, EECS,Peking University, China, where he is currently anAssistant Professor. He published over 30 technicalarticles in refereed journals and proceedings in the

areas of image and video coding, video processing, video streaming, andtransmission.

Ronggang Wang received the B.S. degree in in-formation science from HeiLongJiang University,China, in 1998, the M.S. degree in computer engi-neering from the Harbin Institute of Technology,China, in 2002, and the Ph.D. degree in computerengineering from the Institute of Computing Tech-nology, Chinese Academy of Sciences, in 2006.

Since 2006, he has been a researcher with FranceTelecom R&D Beijing, China. His research interestsare video compression, video streaming, and multi-media processing.

Wen Gao (F’08) received the Ph.D. degree in elec-tronics engineering from the University of Tokyo,Tokyo, Japan.

He is a Professor of computer science at PekingUniversity, China. Before joining Peking University,he was a full Professor of computer science at theHarbin Institute of Technology from 1991 to 1995,and with CAS (Chinese Academy of Sciences), from1996 to 2005, he served as a Professor, ManagingDirector of the Institute of Computing Technologyof CAS, the executive Vice President of the Grad-

uate School of CAS, and a Vice President of the University of Science andTechnology of China. He has published extensively, including four books andover 500 technical articles in refereed journals and conference proceedingsin the areas of image processing, video coding and communication, patternrecognition, multimedia information retrieval, multimodal interface, andbioinformatics.

Dr. Gao is the Editor-in-Chief of the Journal of Computer (a journal of Chi-nese Computer Federation); an Associate Editor of the IEEE TRANSACTIONS ON

CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, the IEEE TRANSACTIONS

ON MULTIMEDIA, and the IEEE TRANSACTIONS ON AUTONOMOUS MENTAL

DEVELOPMENT; an area editor of the EURASIP Journal of Image Communi-cations; and an Editor of the Journal of Visual Communication and ImageRepresentation. He chaired a number of prestigious international conferenceson multimedia and video signal processing and also served on the advisory andtechnical committees of numerous professional organizations.

Authorized licensed use limited to: Harbin Institute of Technology. Downloaded on May 27,2010 at 03:18:32 UTC from IEEE Xplore. Restrictions apply.