Open Access
20 December 2013 Spatiotemporal deinterlacing using a maximum a posteriori estimator based on multiple-field registration
Ho-Taek Lee, Dong-bok Lee, Byungju Lee, Byung Cheol Song
Author Affiliations +
Abstract
This paper proposes an accurate deinterlacing algorithm using a maximum a posteriori (MAP) estimator. First, we produce accurate motion vector fields between the current field and adjacent fields by employing an advanced motion compensation scheme that is suitable for an interlaced format. Next, the progressive frame corresponding to the current field is found via the MAP estimator based on the derived motion vector fields. Here, in order to obtain a stable solution, well-known bilateral total variation–based regularization is applied. Then, at a specific mode decision step, it is decided whether the result from the aforementioned temporal deinterlacing is acceptable or not. Finally, if the temporal deinterlacing is determined to be inappropriate by the mode decision, a typical spatial deinterlacing is applied instead of the MAP estimator-based temporal deinterlacing. Experimental results show that the proposed algorithm provides at maximum 2 dB higher PSNR than a cutting-edge deinterlacing algorithm, while providing better visual quality than the latter.

1.

Introduction

Deinterlacing is an important technique because it converts interlaced video sequences into progressive sequences for progressive digital display devices, such as LCDs, plasma display panels, and organic light-emitting diode TVs. However, visually annoying artifacts, such as edge flicker, jagging, blurring, and feathering, often occur due to imperfect deinterlacing. Thus, a critical issue related to deinterlacing is removing such artifacts as much as possible.

During the past several decades, numerous deinterlacing algorithms have been developed.1 We can categorize the existing algorithms into three groups as follows: spatial deinterlacing,28 motion-adaptive (MA) deinterlacing,911 and motion compensation (MC)-based deinterlacing.1222 Spatial deinterlacing algorithms have the advantages of simple operation and straightforward integration into hardware. The edge-based line averaging (ELA) algorithm is one of the most popular deinterlacing algorithms in this category. However, since all the spatial deinterlacing schemes interpolate the missing pixels by using only intrafield information, the interpolated pixels in the texture area tend to be somewhat blurred.

Primitive temporal approaches such as MA deinterlacing algorithms have been devised to overcome the drawbacks of spatial deinterlacing algorithms by using temporal correlation.911 MA deinterlacing methods detect the existence of motion and then use either an intrafield method, i.e., a typical spatial deinterlacing, or an interfield method. Here, the interield method is a simple interfield interpolation without motion compensation. However, this MA deinterlacing method still suffers from blur or jagging artifacts because it should apply the intrafield method when interpolating missing pixels in moving objects.

Recently, more advanced temporal deinterlacing methods that are based on some form of motion-compensated techniques have become popular.1322 For example, Chang et al. presented an adaptive four-field global/local motion-compensated approach where the same parity four-field motion detection and four-field motion estimation detect static areas and fast motion by four reference fields, and global motion estimation detects camera panning and zooming motions.13 However, this algorithm may give rise to feathering artifacts when the assumption of strong continuity of local motion does not work. Fan and Chung proposed a temporal deinterlacing algorithm using strong spatial-temporal correlation-assisted motion estimation.17 To our knowledge, Fan’s algorithm is the state-of-the-art deinterlacing scheme. However, since conventional temporal deinterlacing algorithms inherently cannot compensate for registration error between the current field and its reference fields, they have a limitation in further improving visual quality.

On the other hand, maximum a posteriori (MAP) estimation has recently been applied to numerous image reconstruction techniques, e.g., deblur, denoising, and superresolution. In statistics, MAP estimation is a method of estimating the parameters of a statistical model. When applied to a data set and given a statistical model, an MAP estimator provides estimates for the model’s parameters. For example, MAP estimators are often used for superresolution reconstruction.23,24 Superresolution restoration aims to solve the following problem: given a set of observed low-resolution images, estimate a high-resolution image. The observed low-resolution images are regarded as degraded observations of a real, high-resolution texture. These degradations typically include geometric warping, optical blur, spatial sampling, and noise. Given several such observations, the MAP estimate of the superresolution image may be obtained such that, when reprojected back into the images via a generative imaging model, it minimizes the difference between the actual and predicted observations.24 Note that if accurate motion information of a certain missing pixel is given, the MAP-based superresolution can reconstruct the missing pixel such that it is very close to its original.

We showed that if the MAP estimator is applied to temporal deinterlacing, subjective visual quality as well as objective visual quality can be improved by minimizing inherent registration errors in missing pixels in Ref. 20. In this paper, we propose an advanced MAP-estimator-based deinterlacing algorithm using high-performance MC method and strong mode decision. The proposed algorithm consists of four steps. First, accurate multiple-field registration is performed between the current field and its neighboring fields. Second, the progressive frame corresponding to the current field is reconstructed via an L1-norm-based MAP estimator using the predicted motion fields. Here, in order to obtain an optimal solution, the well-known steepest descent algorithm is employed. Also, bilateral total variation (BTV)-based regularization is applied to obtain a stable solution and preserve edges well. Third, a mode decision module determines whether the result from the aforementioned temporal deinterlacing is acceptable or not. The proposed mode decision relies on three factors: feathering artifacts, registration errors, and motion vector (MV) correlation. Fourth, if the temporal deinterlacing is determined to be inappropriate by the mode decision, a typical spatial deinterlacing based on edge-directional interpolation is applied instead of the MAP estimator-based temporal deinterlacing. Experimental results show that the proposed algorithm obtains at maximum 2 dB higher peak signal-to-noise ratio (PSNR) than the state-of-the-art spatiotemporal deinterlacing algorithm, i.e., Fan’s algorithm,17 while providing better visual quality.

The remainder of this paper is organized as follows. Section 2 describes the proposed deinterlacing algorithm in detail. Section 3 provides intensive experimental results. Finally, we conclude this paper in Sec. 4.

2.

Proposed Algorithm

As illustrated in Fig. 1, the proposed deinterlacing algorithm has a structure similar to a typical spatiotemporal deinterlacing scheme based on a hard decision between spatial deinterlacing and temporal deinterlacing. The main contribution of this paper is that we can reconstruct even registration errors in temporal deinterlacing by employing a MAP estimator. Note that previous temporal deinterlacing methods output motion-compensated pixels only. The proposed algorithm, meanwhile, can produce outstanding visual quality, especially around diagonal edges, while significantly reducing jagging or feathering artifacts in comparison with the state-of-the-art deinterlacing schemes.

Fig. 1

Block diagram of the proposed deinterlacing algorithm.

JEI_22_4_043038_f001.png

The proposed algorithm consists of four steps, as seen in Fig. 1. First, an advanced motion estimation (ME) algorithm, which is an improved version of the spatial-temporal correlation-assisted search (STFS) proposed in Ref. 17, is applied to the current field and adjacent fields. Second, based on the estimated motion information, MAP estimator-based temporal deinterlacing for the current field is performed. Third, on a block basis, a mode decision module determines whether the result from the temporal deinterlacing is acceptable or not. Fourth, if the temporal deinterlacing result is determined to be unacceptable, a typical spatial deinterlacing is applied. The following subsections describe each step of the proposed algorithm in more detail.

2.1.

Registration Using the Advanced STFS

2.1.1.

Conventional STFS

Prior to describing the advanced STFS, we explain the conventional STFS (Ref. 17) in detail as follows. As illustrated in Fig. 2(a), three consecutive fields, i.e., fn1,fn,fn+1, are involved in the estimation of the motion trajectory. The bidirectional ME and the two kinds of unidirectional ME, i.e., forward and backward ME, are combined in order to fully exploit the information from the three fields.

Fig. 2

Conventional spatial-temporal correlation-assisted search (STFS) scheme.

JEI_22_4_043038_f002.png

Let SADP and SADN be the sum of absolute differences (SAD) between two blocks used in the backward ME and forward ME, respectively, and let SADB be the SAD between two blocks used in the bidirectional ME, where these terms are defined as follows:

Eq. (1)

SADB(q)=(x,y)b|fn+1(x+m,y+n)fn1(xm,yn)|SADN(q)=(x,y)b|fn(x,y)fn+1(x+m,y+n)|SADP(q)=(x,y)b|fn(x,y)fn1(xm,yn)|,
where (x,y) denotes a position in the current block b, and q(m,n) indicates a candidate MV within the given search range. The best MV q^=(m^,n^) for the block b is then obtained by minimizing the following cost:

Eq. (2)

q^=argminq[SAD(q)],
where

Eq. (3)

SAD(q)=SADB(q)+SADN(q)+SADP(q).

Note that the information from the three consecutive fields is involved in the estimation of the motion trajectory. From Fig. 2(a), q^ is chosen as the initial MV candidate vini. For each matching block, there are four spatial MV candidates and five temporal MV candidates, as in Fig. 2(b). The spatial/temporal bidirectional MV candidates are sorted according to the number of occurrences. For the current block, the matching error corresponding to each spatial/temporal MV candidate vST is compared to the matching error of the initial MV candidate vini according to Eq. (4). The comparison is performed starting from the candidate with the maximum number of occurrences.

Eq. (4)

|SAD(vST)SAD(vini)|<T1,
where T1 is a predetermined threshold that is a function of SAD(vini). If Eq. (4) is satisfied, vST is chosen as the final MV. Otherwise, the candidate with the second largest number of occurrences is compared to the initial candidate, and so on. If all the spatial/temporal MV candidates fail to pass the matching error check, vini is chosen as the final MV.

2.1.2.

Advanced STFS

In order to obtain more accurate and denser MVs than the conventional STFS, we apply the overlapped motion compensation scheme to the STFS. Note that MV estimation is performed on an 8×8 block basis, but the matching block size is 16×16. As shown in Fig. 3, the best MV for a specific 8×8 block is obtained by block matching for the extended 16×16 block with the 8×8 block as a center. The conventional STFS is used for this block matching. Finally, the estimated MV for the extended 16×16 block is allocated to the 8×8 center of the matching block. Table 1 shows the superiority of the advanced STFS to the conventional STFS in terms of PSNR. For this experiment, we employed the same progressive common intermediate format (CIF) sequences as in Sec. 3 and produced their interlaced versions in the same fashion as described in Sec. 3. For each method, we computed the PSNR between the motion-compensated frame and the original frame. We compared the averaged PSNR for the first 300 frames of each sequence as an objective evaluation measure. For example, Table 1 shows that the advanced STFS outperforms the conventional STFS by 0.8 dB at maximum for the container sequence. This is because the advanced STFS can produce more accurate and denser MVs than the conventional STFS. Thus, we applied this advanced STFS for registration of the proposed temporal deinterlacing, which is described in the following subsection.

Fig. 3

The overlapped matching of the advanced STFS.

JEI_22_4_043038_f003.png

Table 1

Peak signal-to-noise ratio (PSNR) comparison between conventional spatial-temporal correlation-assisted search (STFS) and the advanced STFS (dB).

NameConventional STFSAdvanced STFS
Foreman37.6938.02
Mobile28.7728.68
M & D47.3747.47
Stefan31.0731.23
Coastguard38.2538.32
Table tennis35.6236.01
Container46.5947.42
Average37.9138.16

2.2.

MAP Estimator-Based Temporal Deinterlacing

The key point of the proposed algorithm is to reconstruct even registration errors in temporal deinterlacing by employing a MAP estimator, whereas previous temporal deinterlacing methods output motion-compensated pixels only. In order to formulate the MAP estimation problem for temporal deinterlacing, we should define the acquisition model of an interlaced field. First, we assume that a progressive frame F may have a motion model M with adjacent frames. Next, a space-invariant point spread function (PSF) H is applied to F for antialiasing. Finally, an interlaced field f is generated by applying an interlace decimation operator D and adding a noise component V. Note that the decimation operator D alternatively subsamples even lines and odd lines, and the decimation operation is defined as follows:

Eq. (5)

fn(x,y)={Fn(x,y),ifxmod2=nmod20,otherwise.

According to Eq. (5), the n’th field fn is derived from the n’th frame Fn. The current field and its neighbor fields are then defined as

Eq. (6)

fn=DnHnMnF+Vnn=1,,N.

Let the processing block size be b×b and the scaling ratio be set to r:1. Then, F in Eq. (6) has a dimension of [r2b2×1], which is arranged in lexicographic order. The [r2b2×r2b2] matrix Mn is the geometric motion operator between the original progressive frame F and the n’th interlaced field fn of size [b2×1]. The PSF is modeled by the [r2b2×r2b2] matrix H. The [b2×r2b2] matrix Dn represents the decimation operator and the [b2×1] matrix Vn denotes the system noise. Based on this model, we propose the MAP estimator-based temporal deinterlacing algorithm.

In general, a MAP estimator for the original progressive frame F given the observed interlaced fields fn can be formulated to find the MAP estimate F^ via the following Lp minimization:

Eq. (7)

F^=argminF[n=1NDnHMnFfnpp+λΓ(F)],
where λ is a control parameter and N is the number of reference fields for MAP estimation. Note that Hn=H because a common PSF is assumed for all fields without loss of generality. The second term is known as regularization. Consider regularization in deinterlacing is very useful for finding a stable solution.23 Among many possible regularization terms, we select BTV-based regularization, which results in a full progressive frame with sharp edges and is easy to implement. The BTV-based regularization term is defined by

Eq. (8)

l=ττm=0l+m0τα|l|+|m|FSxlSymF1,
where τ and α are parameters to control the regularization level, and Sxl and Sym are operators to shift by l and m pixels in the horizontal and vertical directions, respectively. This paper employs L1-norm regularization owing to its simplicity and edge-preserving property. Also, L1-norm is applied to the data term, i.e., the first term of Eq. (7), to minimize sensitivity to noise, i.e., p=1.

In addition, we compare BTV-based regularization of Eq. (8) with a typical TV-based regularization23 to show the superiority of the BTV-based regularization quantitatively and qualitatively. Table 2 compares BTV-based regularization and TV-based regularization in terms of PSNR. The PSNR value in Table 2 is the average PSNR for the first 50 deinterlaced frames per sequence. We can see that BTV-based regularization provides at maximum 1.1 dB and on average 0.4 dB higher PSNRs than TV-based regularization. For example, Fig. 4 shows the deinterlacing results for carphone and foreman sequences. Note that the TV-based regularization causes some artifacts for flat areas due to its sensitivity to noise. On the other hand, the proposed BTV-based regularization seldom shows such a phenomenon because it is inherently robust against noise and preserves the details such as edges and textures better than TV-based regularization. As a result, we can achieve high-quality temporal deinterlacing via this robust regularized MAP estimator.

Table 2

PSNR comparison between total variation (TV) regularization and bilateral total variation (BTV) regularization (dB).

NameTV regularizationBTV regularization
Foreman37.7538.87
Stefan31.8932.03
Carphone33.9434.15
Flower31.2031.61
Football33.0833.10
Highway36.3936.73
Salesman37.9638.61
Average34.6035.01

Fig. 4

The deinterlaced frames according to regularization term.

JEI_22_4_043038_f004.png

We use the steepest descent algorithm to find the solution to Eq. (7). We can thus obtain an optimal solution in the following iterative manner. That is, the (t+1)’th MAP estimate F^t+1 is derived from the t’the estimate F^t.

Eq. (9)

F^t+1=F^tβ{n=1NMnTHTDnTsign(DnHMnF^tfn)+λl=ττm=0l+m0τα|l|+|m|[ISymSxl]sign(F^tSxlSymF^t)},
where β is a scalar defining the step size in the direction of the gradient and I is an identity matrix. The matrices M, H, and D and their transposes can be exactly interpreted as direct image operators, such as shift, blur, and decimation.23 Noting and implementing the effects of these matrices as a sequence of operators spares us from explicitly constructing them as matrices. This property allows the proposed temporal deinterlacing method to be implemented in an extremely fast and memory efficient way. After several iterations according to Eq. (9), an optimal progressive frame F^ corresponding to the current field is reconstructed from N consecutive fields including the current field. Finally, MAP estimator-based temporal deinterlacing is completed by replacing the missing field pixels in the current field with the corresponding pixels in F^.

2.3.

Mode Decision

Conventional spatiotemporal deinterlacing methods often suffer from feathering, blur, and jagging artifacts caused by false mode decision. For instance, feathering artifacts occur near object boundaries due to occlusion or inaccurate MVs. Mode decision, hence, significantly affects the visual quality of spatiotemporal deinterlacing algorithms. Fan and Chung employed the so-called slope detector (SD)-based feathering artifact detector due to its low complexity and high detection ability.17

This paper presents a stronger mode decision method that additionally takes into account mean of absolute differences (MAD) values, MV correlation, and MAD correlation. Figure 5 illustrates the proposed mode decision scheme where the MV reliability is examined on a block basis in two steps. Assume that the MVs are produced by the advanced STFS. First, for each block that is motion-compensated by the advanced STFS, feathering artifacts are detected by the aforementioned SD-based detector of Ref. 17. If a feathering artifact is not detected, temporal deinterlacing is applied to the block. Otherwise, additional examination is performed with more information. The reason why we chose such a conservative approach is that the blur phenomenon caused by spatial deinterlacing is visually less annoying than feathering artifacts. The second examination procedure is explained in greater detail below.

Fig. 5

The proposed mode decision scheme.

JEI_22_4_043038_f005.png

Since the SD-based detector depends on the motion-compensated image only, it may cause false positives in detecting various feathering artifacts. We hence propose an additional step to reduce the false positive rate of the first step by taking into account MAD values, MV correlation, and MAD correlation. As shown in Fig. 5, if the MAD of the current block is sufficiently small, i.e., MAD(v^)<δ1, and the MAD and MV of the current block are highly correlated with those of its neighbors, i.e., cor(MV)<δ2&cor(MAD)<δ3, we determine that the current block has a reliable MV. Let cor(MV) be the L1-norm distance between the current MV and the component-wise median of MVs of its eight-connected blocks, while cor(MAD) indicates the L1-norm distance between MAD(v^) and the average MAD for the eight-connected neighbor blocks. In this paper, δ1, δ2, and δ3 are empirically set to 20, 5, and 5, respectively. Therefore, since this second decision step further examines the MV reliability independently of the motion-compensated image, it can effectively prevent the entire mode decision from being trapped in false positives.

2.4.

Spatial Deinterlacing

For spatial deinterlacing of the proposed algorithm, we adopted a block-based directional interpolation algorithm proposed by Chen and Tai in Ref. 14. This subsection briefly introduces Chen’s algorithm. Figure 6 shows a deinterlaced frame, where the dotted line indicates the missing scan line that needs to be interpolated and the solid lines indicate the original field data. It is assumed that the missing pixel fn(x,y) is centered in a target block BC; BUq and BLq are defined as the corresponding upper and lower referenced blocks, respectively, and q is assumed to be a candidate directional vector between the BC and its upper candidate block. Note that the blocks BC, BUq, and BLq only contain pixels in fn. The best directional vector q^ is detected using the following equation:

Eq. (10)

q^=argminqΩ[SADU(q)+SADL(q)],
where SADU(q) and SADL(q) denote the SAD between BC and BUq, and the SAD between BC and BLq, respectively. The search range of q, i.e., Ω, includes 23 directional vectors so as to detect edges whose degrees are greater than 7.5. Let SAD(q^) be the minimal cost according to Eq. (10). When SAD(q^) is greater than a certain threshold θ1, or the difference between SAD(q^) and SAD(q^+90deg) is not greater than a threshold θ2, the selected q^ is not reliable, and the typical line average filter is used to construct the final result. In this paper, θ1 and θ2 are empirically set to 30 and 10, respectively.

Fig. 6

Edge-based spatial deinterlacing.

JEI_22_4_043038_f006.png

3.

Experimental Results

3.1.

Experimental Condition

First, for an objective quality evaluation, we used 12 progressive CIF sequences; foreman, mobile, mother and daughter (M&D), Stefan, coastguard, table tennis, container, flower garden, football, highway, Paris, and tempete. We generated the interlaced fields by applying a simple low-pass filter of {1, 2, 1} to those progressive frames and alternatively subsampling the low-pass filtered frames. After deinterlacing, we computed the PSNR between the reconstructed progressive frame and the original frame. Note that only Y components are treated here. We chose the averaged PSNR for the first 300 frames of each sequence as an objective evaluation measure.

We compared the proposed algorithm with LA, ELA, Lee’s algorithm,9 Chang’s algorithm,13 Chen’s algorithm,14 Fan’s algorithm,17 Yang’s algorithm,18 Mohammadi’s algorithm,21 and Trocan’s algorithm.22 The number of reference fields for the proposed temporal deinterlacing was set to 3, i.e., the current field and its temporally previous and next fields.

The search range for registration in the proposed algorithm and Chang’s algorithm was ±32 both horizontally and vertically. All the block sizes for matching in the advanced STFS were set to 16×16 and the block sizes of overlapping are set to 8×8. In Eq. (8), α and τ are fixed to 0.5 and 2, respectively. Figure 7 shows the influence of a parameter α on the overall performance of the proposed algorithm for four test sequences, i.e., highway, carphone, foreman, and football. For this experiment, when deinterlacing the first field of each sequence, we progressively changed α from 0.1 to 1.0 and tracked the PSNR values accordingly. From Fig. 7, we can observe that as α becomes larger than 0.5, the PSNR starts to decrease. So, we set α to 0.5. Through a similar experiment, we found that another parameter τ rarely affects the overall performance of the proposed algorithm. So, we fixed τ to an acceptable value. Table 3 shows the appropriate values of the other parameters, i.e., β and λ chosen for each sequence to maximize the temporal deinterlacing performance in terms of PSNR. Note that β and λ should be determined on a shot basis in a sequence. For instance, since the foreman and table tennis sequences consist of two different shots, we derived and applied two different parameters to those sequences, as given in Table 3.

Fig. 7

Peak signal-to-noise ratio (PSNR) performance according to α for four video sequences.

JEI_22_4_043038_f007.png

Table 3

Parameter setting for temporal deinterlacing.

NameShot #1Shot #2Iteration
βλβλ
Foreman1.30.21.50.15
Mobile1.10.054
M&D0.50.24
Stefan1.00.32
Coastguard0.20.24
Table tennis0.50.20.70.15
Container0.10.24
Flower garden1.50.0550
Football1.10.0510
Highway0.90.1510
Paris1.50.0530
Tempete1.50.0510

Figure 8 shows the effects of the parameters in Eq. (7) on the overall performance of the proposed temporal deinterlacing. This experiment was performed for a field in the first shot of the foreman sequence. From Fig. 8(a), we can observe that if the optimal parameters are used, the PSNR performance is very stable. On the contrary, Fig. 8(b) shows that if nonoptimal parameters, i.e., the parameters of the second shot, are employed, the PSNR performance drastically decreases as the iterations go on, and the peak PSNR also becomes low in comparison with that in Fig. 8(a). From Fig. 8, we can also find that the termination point of iteration is important. Thus, we selected the optimal number of iterations for each sequence, as given in Table 3.

Fig. 8

PSNR values according to the number of iterations for the first shot of the foreman sequence. (a) Results with the optimal parameters. (b) Results with nonoptimal parameters.

JEI_22_4_043038_f008.png

3.2.

Performance Evaluation

Table 4 shows the PSNR results of several algorithms for 12 test sequences. For this experiment, we implemented LA, ELA, Chang’s algorithm,13 and Mohammadi’s algorithm21 personally and verified them. We can see that the proposed algorithm provides ~5dB higher PSNR than Chang’s13 and Mohammadi’s algorithms21 on average. Especially, for container sequence, the proposed algorithm goes beyond Chang’s algorithm13 up to 12 dB.

Table 4

PSNR comparison for different deinterlacing algorithms (dB).

SequencesLine averaging (LA)Edge-based LA (ELA)Ref. 13Ref. 21Proposed
Foreman33.634.234.433.038.6
Mobile25.423.526.824.833.9
M&D39.838.844.641.646.2
Stefan28.327.329.728.532.4
Coastguard28.527.830.331.037.1
Table tennis28.927.833.531.935.3
Container28.927.934.738.346.7
Flower22.622.122.827.930.5
Football34.834.432.731.435.7
Highway32.332.432.531.437.0
Paris27.126.833.233.535.3
Tempete29.228.329.928.932.3
Average30.029.332.131.836.8

In addition, Table 5 provides PSNR comparison results for several recent methods in the literature. In Table 5, the numerical values of Refs. 9, 14, 17, 18, and 22 were directly extracted from those papers. Hence, a few values are not available in the table. Note that the proposed algorithm provides on average 0.8 dB higher PSNR than the cutting-edge algorithm, i.e., Fan’s algorithm.17 For the Stefan sequence, in particular, we achieve a significant PSNR improvement of ~2dB over Fan’s algorithm. We also find from the available PSNR values in Table 5 that the proposed algorithm is superior to Trocan’s algorithm.22 As a result, the proposed algorithm outperforms the previous works in terms of objective visual quality of PSNR. In addition, Fig. 9 shows the PSNR values according to the frame number. We can state that the proposed algorithm consistently provides higher PSNR values than the previous works.

Table 5

PSNR comparison for different deinterlacing algorithms in the literature (dB).

SequencesRef. 9Ref. 14Ref. 17Ref. 18Ref. 22Proposed
ForemanN/A33.838.132.937.338.6
Mobile30.327.733.3N/A31.533.9
M&D46.245.045.8N/AN/A46.2
Stefan28.028.830.526.531.332.4
Coastguard32.9N/A35.930.9N/A37.1
Table tennis34.037.036.428.2N/A35.3
Container39.645.545.344.2N/A46.7
Flower gardenN/AN/AN/AN/AN/A30.5
FootballN/AN/AN/AN/AN/A35.7
HighwayN/AN/AN/AN/AN/A37.0
ParisN/AN/AN/AN/AN/A35.3
TempeteN/AN/AN/AN/AN/A32.3
Note: Bold values indicate the maximum PSNR for each sequence.

Fig. 9

PSNR performance according to the frame number. (a) Foreman. (b) Mobile. (c) Container.

JEI_22_4_043038_f009.png

In Figs. 10 and 11, the deinterlaced results for mobile and Stefan sequences are compared. Note that the proposed algorithm produces a deinterlaced result close to its original frame, while Chang’s algorithm and LA suffer from visually annoying artifacts, such as jagging and blur. For the mobile sequence (see Fig. 10), we find that the proposed algorithm outperforms the others. Especially observing the numbers in the calendar, the proposed algorithm shows almost the same visual quality as the original frame. From Fig. 11, we can find a perfect court line generated by the proposed temporal deinterlacing. Note that the competitors show severe jagging artifacts and line-crawling in the near horizontal edges.

Fig. 10

The deinterlaced frames for the mobile sequence. (a) Original. (b) Line averaging (LA). (c) Chang’s method.13 (d) Proposed method.

JEI_22_4_043038_f010.png

Fig. 11

The deinterlaced frames for the Stefan sequence. (a) Original. (b) LA. (c) Chang’s method.13 (d) Proposed method.

JEI_22_4_043038_f011.png

In addition, Figs. 12 and 13 show the deinterlaced results for real interlaced video sequences. We adopted two well-known interlaced sequences (720×480i), car and table tennis. Even for real interlaced video sequences, the proposed algorithm still provides outstanding visual quality without any artifacts in comparison with the existing methods. For example, LA and Chang’s algorithm show jagging and feathering artifacts [see Figs. 12(b) and 12(c)], but the proposed algorithm does not cause such artifacts, as shown in Fig. 12(d).

Fig. 12

The results for a cropped region of the car sequence. (a) Input interlaced frame. (b) LA. (c) Chang’s method.13 (d) Proposed method.

JEI_22_4_043038_f012.png

Fig. 13

The results for a cropped region of the table tennis sequence. (a) Input interlaced frame. (b) LA. (c) Chang’s method.13 (d) Proposed method.

JEI_22_4_043038_f013.png

In addition, we measured the execution times of several algorithms. This experiment was executed on a quad-core CPU at 2.66 GHz with 3 GB DDR2 DRAM. Since motion estimation occupies most of the complexity of MC-based deinterlacing algorithm, we adopted a fast full search25 and a famous fast search algorithm, i.e., enhanced predictive zonal search (EPZS)26 for fast motion estimation in this experiment. Table 6 compares the proposed algorithm with LA, ELA, and Ref. 13 in terms of the CPU running time. Note that each numerical value indicates the average for the first 10 fields of foreman sequence, and all the algorithms were implemented in MATLAB®. We could find that in comparison with fast full search, the EPZS cuts the CPU running time of the proposed algorithm in half with an acceptable PSNR drop of ~1dB on average. The proposed algorithm using EZPS still provides ~1.8 times longer execution time than Ref. 13. However, if we employ several optimization skills additionally, we can significantly reduce the computational cost of the proposed algorithm.

Table 6

CPU running times for various algorithms (s).

LAELARef. 13Proposed (fast full search)Proposed (enhanced predictive zonal search)
0.0050.099.634.517.4

4.

Concluding Remarks

This paper presents a robust temporal deinterlacing algorithm based on an MAP estimator. First, registration using an advanced STFS algorithm is performed between the current field and its neighboring fields. Second, the progressive frame corresponding to the current field is found via an L1-norm-based MAP estimator based on the predicted interfield MV information. Third, a mode decision module determines whether the result from the temporal deinterlacing is acceptable or not. Finally, edge-directional interpolation is applied to the pixels whose MVs are not reliable, instead of the aforementioned temporal deinterlacing. Experimental results show that the proposed algorithm yields at maximum 2 dB higher PSNR than the cutting-edge deinterlacing algorithm,17 while providing better visual quality.

Acknowledgements

This work was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (2012000446), and was supported by the MSIP (Ministry of Science, ICT & Future Planning), Korea, under IT/SW Creative research program supervised by the NIPA (National IT Industry Promotion Agency) (NIPA-2013-H0502-13-1010), and supported by INHA UNIVERSITY Research Grant.

References

1. 

G. de HaanE. B. Bellers, “Deinterlacing: an overview,” Proc. IEEE, 86 (9), 1839 –1857 (1998). http://dx.doi.org/10.1109/5.705528 IEEPAD 0018-9219 Google Scholar

2. 

M. K. Parket al., “New edge dependent de-interlacing algorithm based on horizontal edge pattern,” IEEE Trans. Consum. Electron., 49 (4), 1508 –1512 (2003). http://dx.doi.org/10.1109/TCE.2003.1261260 ITCEDA 0098-3063 Google Scholar

3. 

M. K. Parket al., “Edge-dependent interpolation-based de-interlacing using edge patterns,” J. Electron. Imaging, 15 (4), 043003 (2006). http://dx.doi.org/10.1117/1.2400850 JEIME5 1017-9909 Google Scholar

4. 

J. WangG. JeonJ. Jeong, “Deinterlacing algorithm with an advanced non-local means filter,” Opt. Eng., 51 (4), 047009 (2012). http://dx.doi.org/10.1117/1.OE.51.4.047009 OPEGAR 0091-3286 Google Scholar

5. 

H. Y. Leeet al., “ Adaptive scan rate up-conversion system based on human visual characteristics,” IEEE Trans. Consum. Electron., 46 (4), 999 –1006 (2000). http://dx.doi.org/10.1109/30.920453 ITCEDA 0098-3063 Google Scholar

6. 

J. LeeC. Lee, “Deinterlacing with reduced flickering,” Opt. Eng., 51 (6), 067401 (2012). http://dx.doi.org/10.1117/1.OE.51.6.067401 OPEGAR 0091-3286 Google Scholar

7. 

Y. T. Kim, “De-interlacing algorithm based on sparse wide vector correlations,” Proc. SPIE, 2727 89 –99 (1996). http://dx.doi.org/10.1117/12.233311 PSISDG 0277-786X Google Scholar

8. 

W. KimS. JinJ. Jeong, “Novel intra deinterlacing algorithm using content adaptive interpolation,” IEEE Trans. Consum. Electron., 53 (3), 1036 –1043 (2007). http://dx.doi.org/10.1109/TCE.2007.4341583 ITCEDA 0098-3063 Google Scholar

9. 

K. LeeJ. LeeC. Lee, “Deinterlacing with motion adaptive vertical temporal filtering,” IEEE Trans. Consum. Electron., 55 (2), 636 –643 (2009). http://dx.doi.org/10.1109/TCE.2009.5174433 ITCEDA 0098-3063 Google Scholar

10. 

S. F. LinY. L. ChangL. G. Chen, “Motion adaptive interpolation with horizontal motion detection for deinterlacing,” IEEE Trans. Consum. Electron., 49 (4), 1256 –1264 (2003). http://dx.doi.org/10.1109/TCE.2003.1261227 ITCEDA 0098-3063 Google Scholar

11. 

S. F. LinY. L. ChangL. G. Chen, “Motion adaptive interpolation with morphological operation and 3:2 pull-downed recovery for de-interlacing,” in Proc. IEEE Int. Conf. on Multimedia and Expo, 661 –664 (2002). Google Scholar

12. 

H. ChoiC. Lee, “Motion adaptive deinterlacing with modular neural networks,” IEEE Trans. Circ. Syst. Video Technol., 21 (6), 844 –849 (2011). http://dx.doi.org/10.1109/TCSVT.2011.2129190 ITCTEM 1051-8215 Google Scholar

13. 

Y. L. Changet al., “Video de-interlacing by adaptive 4-field global/local motion compensated approach,” IEEE Trans. Circ. Syst. Video Technol., 15 (12), 1569 –1582 (2005). http://dx.doi.org/10.1109/TCSVT.2005.858746 ITCTEM 1051-8215 Google Scholar

14. 

Y. R. ChenS. C. Tai, “True motion-compensated de-interlacing algorithm,” IEEE Trans. Circ. Syst. Video Technol., 19 (10), 1489 –1498 (2009). http://dx.doi.org/10.1109/TCSVT.2009.2022782 ITCTEM 1051-8215 Google Scholar

15. 

J. Changet al., “Adaptive arbitration of intra-field and motion compensation methods for de-interlacing,” IEEE Trans. Circ. Syst. Video Technol., 19 (8), 1214 –1220 (2009). http://dx.doi.org/10.1109/TCSVT.2009.2020341 ITCTEM 1051-8215 Google Scholar

16. 

Q. Huanget al., “Deinterlacing using hierarchical motion analysis,” IEEE Trans. Circ. Syst. Video Technol., 20 (5), 673 –686 (2010). http://dx.doi.org/10.1109/TCSVT.2010.2045807 ITCTEM 1051-8215 Google Scholar

17. 

Y. C. FanC. H. Chung, “De-interlacing algorithm using spatial-temporal correlation-assisted motion estimation,” IEEE Trans. Circ. Syst. Video Technol., 19 (7), 932 –944 (2009). http://dx.doi.org/10.1109/TCSVT.2009.2020327 ITCTEM 1051-8215 Google Scholar

18. 

S. Yanget al., “Motion compensation assisted motion adaptive interlaced-to-progressive conversion,” IEEE Trans. Circ. Syst. Video Technol., 14 (9), 1138 –1148 (2004). http://dx.doi.org/10.1109/TCSVT.2004.833163 ITCTEM 1051-8215 Google Scholar

19. 

J. DongK. N. Ngan, “Real-time de-interlacing for H.264-coded HD videos,” IEEE Trans. Circ. Syst. Video Technol., 20 (8), 1144 –1149 (2010). http://dx.doi.org/10.1109/TCSVT.2010.2056952 ITCTEM 1051-8215 Google Scholar

20. 

H. LeeT. H. LeeB. C. Song, “Spatio-temporal de-interlacing based on maximum likelihood estimation,” in IEEE Visual Communications and Image Processing, 1 –4 (2011). Google Scholar

21. 

H. M. MohammadiY. SavariaJ. M. P. Langlois, “Enhanced motion compensated deinterlacing algorithm,” IET Image Process., 6 (8), 1041 –1048 (2012). http://dx.doi.org/10.1049/iet-ipr.2011.0124 1751-9659 Google Scholar

22. 

M. TrocanB. MikovicovaD. Zhanguzin, “An adaptive motion-compensated approach for video deinterlacing,” Multimed. Tools Appl., 61 (3), 819 –837 (2012). http://dx.doi.org/10.1007/s11042-011-0845-7 1380-7501 Google Scholar

23. 

S. Farsiuet al., “Fast and robust multiframe super-resolution,” IEEE Trans. Image Process., 13 (10), 1327 –1344 (2004). http://dx.doi.org/10.1109/TIP.2004.834669 IIPRE4 1057-7149 Google Scholar

24. 

D. CapelA. Zisserman, “Super-resolution from multiple views using learnt image models,” in Proc. of 2001 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, II-627 –II-634 (2001). Google Scholar

25. 

M. BrunigW. Niehsen, “Fast full-search block matching,” IEEE Trans. Circ. Syst. Video Technol., 11 (2), 241 –247 (2001). http://dx.doi.org/10.1109/76.905989 ITCTEM 1051-8215 Google Scholar

26. 

A. M. Tourapis, “Enhanced predictive zonal search for single and multiple frame motion estimation,” in Proc. Visual Communications and Image Processing, 1069 –1079 (2002). Google Scholar

Biography

Ho-Taek Lee received his BS and MS degrees in electronic engineering from Inha University, Incheon, Korea, in 2010 and 2012, respectively. Currently, he is working for LG Electronics in Seoul, Republic of Korea. His research interests include image processing and video coding.

Dong-bok Lee received his BS and MS degrees in electronic engineering from Inha University, Incheon, Korea, in 2011 and 2013, respectively. Currently, he is pursuing his PhD degree in Inha University. His research interests include image processing and video coding.

Byungju Lee received his BS degree in electronic engineering from Inha University, Incheon, Korea, in 2012. Currently, he is pursuing his MS degree in Inha University. His research interests include image processing and video coding.

Byung Cheol Song received his BS, MS, and PhD degrees in electrical engineering from the KAIST, Daejeon, Korea, in 1994, 1996, and 2001, respectively. From 2001 to 2008, he was a senior engineer at Digital Media R&D Center, Samsung Electronics Co., Ltd., Suwon, Korea. In March 2008, he joined the School of Electronic Engineering, Inha University, Incheon, Korea, and currently is an associate professor. His research interests are in the general areas of image/video processing.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Ho-Taek Lee, Dong-bok Lee, Byungju Lee, and Byung Cheol Song "Spatiotemporal deinterlacing using a maximum a posteriori estimator based on multiple-field registration," Journal of Electronic Imaging 22(4), 043038 (20 December 2013). https://doi.org/10.1117/1.JEI.22.4.043038
Published: 20 December 2013
Advertisement
Advertisement
KEYWORDS
Visualization

Motion estimation

Fluctuations and noise

Statistical analysis

Error analysis

Super resolution

Algorithm development

Back to Top