I Introduction
Motion estimation (ME) and motion compensation (MC) are the fundamental techniques of video coding to remove the temporal redundancy between video frames. Block matchingbased ME and blockbased MC have been integrated into the reference softwares of almost all the existing video coding standards, including the currently widely adopted H.264/MPEG4 AVC [1] and the stateoftheart H.265/MPEGH High Efficiency Video Coding (HEVC) [2]. The underlying model of blockbased MC is translational motion model, which is too simple to efficiently describe the complex motions in natural videos, such as rotation and zooming. During the development of the video coding standards, many efforts have been made to characterize the complex motions. For example, partitioning blocks into smaller ones can handle complex motions to some extent [3], but may incur more overhead bits for block partitions and more motion vectors (MVs). Therefore, since the translational motion model has not been changed, the standardbased video coding framework is unable to represent the complex motions such as rotation and zooming efficiently.
In the year of as early as 1993, Seferidis et al. [4] pointed out that highorder motion models, such as affine, bilinear, and perspective motion models, were more efficient to characterize the complex motions than the translational one. Among all the highorder motion models introduced in [4], affine motion model has received the most attention of research due to its simplicity. Previous work on affine motion model can be divided into two categories: global affine motion model and local affine motion model.
For global affine motion model [5, 6], usually several groups of model parameters are used to build global affine motion models between two frames, and then each reference frame is warped several times using different model parameters to generate multiple warped references. However, due to a limited number of global affine motion models, such methods are less capable of providing accurate motion parameters for every local motion region. Besides, the increased number of reference frames due to the warped ones will increase the ME computations significantly.
Local affine motion model can be further categorized into meshbased and generalized blockbased. In the meshbased methods [7, 8], the vertices of the mesh are known as control points, whose MVs are used to determine the motions of all the other pixels through locally variant motion models. Since the control point is shared by the neighboring blocks, it is difficult for us to determine the MVs of the control point through ME in a blockbased ratedistortion optimization (RDO) process due to the spatial dependency between neighboring blocks. Therefore, the meshbased methods cannot be well integrated into the modern video coding standards.
In the generalized blockbased methods [9, 10], each block can determine its own affine motion parameters. This is consistent with the standardized video coding framework, only replacing MVs by affine motion parameters for MC. Generalized blockbased affine motion model is intuitively promising to better characterize complex motions, thereby improving coding efficiency. However, both ME and MC under affine motion models are significantly more complex than the traditional block matchingbased ME and blockbased MC. Existing works have not achieved a good balance between coding efficiency and coding complexity. Some of them failed to achieve significantly better ratedistortion (RD) performance, whilst others had too much complexity.
In this paper, we propose a video coding framework using the approach of generalized blockbased affine motion model, which can achieve a better tradeoff between coding efficiency and computational complexity. The framework can be seamlessly integrated into the modern video coding standards, e.g., HEVC. In summary, the proposed framework mainly has the following key contributions:

A fourparameter affine motion model is studied in this paper. Different from previous works that adopted the sixparameter model, the fourparameter model has only four degrees of freedom or equivalently needs only two MVs to represent. It saves two parameters for each block. Meanwhile, this model can accurately characterize rotation, zooming, translation, and any combination of them. Therefore, it can handle most of the complex motions in natural videos.

To efficiently encode affine motion parameters, we propose two techniques: advanced affine motion vector prediction combined with fast affine ME, and affine model merge. The proposed fast affine ME algorithm iteratively updates the two MVs of a block according to gradient descent. It was originally proposed for the sixparameter affine motion model in our previous work [11], and extended for the fourparameter model herein. The proposed gradientbased fast affine ME algorithm can converge very fast, thus can reduce the encoding complexity significantly. In addition, the proposed affine model merge tries to reuse the affine motion parameters of neighboring blocks instead of regenerating a new model from the MVs of the neighbors. The affine model merge can make full use of the motion model correlation between neighboring blocks, and thus can improve the coding efficiency.

We also propose two fast affine MC techniques to reduce both the encoding and decoding complexities. A onestep subpixel interpolation filter which can decrease the interpolation times significantly is developed to replace the previous twostep subpixel interpolation. Moreover, the blockbased MC rather than the pixelbased MC is adopted for acceleration, together with an adaptive choice of block size to ensure interpolation precision.
We perform experiments to verify the efficiency of the proposed video coding framework integrated with HEVC. Compared to HEVC main profile, our proposed techniques altogether can achieve significant bitrate savings while maintain computational efficiency.
This paper is organized as follows. In Section II, we will give a brief review of the related works. The proposed lowcomplexity fourparameter affine motion model based framework will be introduced in Section III. The experimental results are shown in Section IV. Finally, Section V concludes this paper.
Ii Related Work
The affine motion model utilized in ME and MC can be divided into two categories: global affine motion model and local affine motion model. For global affine motion model, Wiegand et al. [5] proposed to use several global affine motion models to generate several warped reference frames. The warped reference frames were used to obtain a better prediction, and the index of the warped reference frame was needed to be transmitted to the decoder. To reduce the overhead bitrate, Li et al. [6] developed a 4D vector quantizer to code the affine motion parameters more efficiently. Besides, Yu et al. [12] proposed to use only one global affine motion model, and the MVs of the salient features between the original frame and reference frame were used to determine the affine motion parameters to generate a warped reference frame. However, the global affine motion model based methods cannot provide accurate affine motion parameters for each local motion region.
The local affine motion model can be utilized in two manners: meshbased methods and generalized blockbased methods. Nakaya and Harashima [7] firstly proposed to use 2D mesh to perform MC and designed a simple method to determine the MVs of the control points. Toklu et al. [13] proposed to add control points hierarchically to better determine the motion models of each block. AlRegib et al. [14] further developed the method and proposed a contentbased irregular mesh to better describe the object boundary. However, due to the various block sizes in modern video coding framework, the problem of determining the MVs of control points through RDO remains difficult.
Besides the meshbased methods, the generalized blockbased methods have also been studied by many researchers. Kordasiewicz et al. [15] considered to derive a better prediction block through affine motion model using the surrounding translational MVs. Besides, Cheung and Siu [9] proposed to use the neighboring information to estimate the affine motion parameters of the current block and added an affine mode into the mode decision process. Then Narroschke and Swoboda [16] found that the affine motion model was more suitable for the large blocks introduced in HEVC. Huang et al. [17] extended the work in [9] for HEVC and designed the affine skip/direct mode to improve the coding efficiency. This work was further developed to a quite complex affine MC framework and many coding modes including affine skip/direct, affine merge, and affine inter were designed to fully exploit the motion correlation between neighboring blocks [10]. Moreover, Heithausen and Vorwerk [18] investigated and compared the performance of different kinds of highorder motion models in HEVC. Also, with the development of the merge mode [19] in HEVC, Chen et al. [20] further developed the affine skip/direct mode to incorporate with the merge mode for the translational motion model and proposed to add some temporal motion candidates into the candidate lists. However, the previous affine merge schemes always attempted to regenerate a new affine motion model through the motion information of the neighboring blocks. Since the neighboring blocks may correspond to different objects or have totally different motion models, the regenerated affine motion model may be inaccurate.
There is a class of local affine motion modeling algorithms designed specifically for the zooming motion in videos. Yuan et al. [21] proposed to use the zooming model to generate a better motion vector predictor (MVP) for the current block. Since this work only generated a better MVP, the RD performance improvement was limited. The algorithm in [22]
further developed a zooming motion model to better characterize the zooming motion and proposed to use linear regression to estimate the MV of the current block from the MVs of the neighboring blocks. Besides, Po
et al. [23] proposed to generate multiple zooming references using a group of model parameters and designed a subsampled blockmatching algorithm to reduce the complexity of ME over a number of reference frames. Kim et al. [24] proposed a 3D diamond pattern search to reduce the number of search points during ME.One critical issue, which hinders the adoption of affine as well as other highorder motion models, is the significant increase of ME complexity. In fact, in the modern video coding framework, the ME process always takes the majority of the encoding time even for translational motion model. Due to the high complexity of ME, the fast ME algorithms [25, 26] have been hot research topics for a long time. For example, the famous Enhanced Predictive Zonal Search (EPZS) [27] algorithm was adopted into the H.264/AVC reference software. HEVC reference software integrated a socalled Test Zone Search (TZS) [28] method which was a further development of EPZS. Both methods show quite good tradeoffs between the RD performance and encoding complexity for the ME of the translational motion model. However, it is not easy to apply them to highorder motion models, for which more parameters need to be determined through ME. Although there were also some algorithms trying to design fast ME algorithm for zooming motion [24], it is not easy to extend those algorithms to more general cases. Therefore, there is an urgent need to design a fast ME algorithm for highorder motion models.
Iii The Proposed Fourparameter Affine MC Framework
The proposed fourparameter affine MC framework will be introduced from three aspects. Firstly, we will introduce the derivation and representation of the proposed fourparameter affine motion model. Secondly, the two methods to encode the affine MVs will be introduced in detail. Thirdly, we will introduce the coding tools to speed up the MC process.
Iiia The fourparameter affine motion model
IiiA1 The derivation of the fourparameter affine motion model
The typical sixparameter affine motion model can be described as
(1) 
where , , , , , and are the six affine motion parameters. The and are the coordinates of the same pixel before and after the transform of the affine motion model. In essence, an affine transform is any transform that preserves lines and parallelism. Therefore, the affine motion model can characterize translation, rotation, zooming, shear mapping, and so on. However, the most common motions in daily videos only include six kinds of typical camera motions (camera track, boom, pan, tilt, zoom, and roll) and the typical object motions (translation, rotation, and zooming), for which six model parameters are more than necessary. It should be noted that the rotation here means the object rotates in a 2D plane that is parallel with the camera. Also, the object zooming can be characterized using an affine motion model only if the relative distance between the object and camera keeps unchanged or the object has a planar surface. Since this paper focuses on the local affine model, we can assume that a local block has a planar surface as long as the block is small enough. In the following, the typical object motions will be used as examples to explain the physical interpretation of the proposed fourparameter affine motion model.
In fact, as shown in Fig.1 (a), if only the combination of rotation and translation is needed to be characterized, the relationship between the coordinates of the same pixel before and after the transformation can be described as
(2) 
where is the rotation angle. Besides, as shown in Fig. 1 (b), if only the combination of zooming and translation is to be characterized, the relationship can be described as
(3) 
where is the zooming factor in both and directions, respectively.
Both the combinations of rotation/zooming and translation need three parameters to characterize. If we combine the rotation, zooming, and translation together, four parameters will be needed and the relationship can be described as
(4) 
If we replace and with and , (4) can be rewritten as
(5) 
where and are the horizontal and vertical components of MV for the position . Eq. (5) is the fourparameter affine motion model used in this paper to accurately characterize the combination of rotation, zooming, and translation.
Comparing the sixparameter affine motion model in (1) with the fourparameter affine motion model in (5), it can be obviously seen that fewer parameters will be calculated under the fourparameter affine motion model for each block thus the decoding complexity can be slightly reduced. Besides, the encoding complexity will also be reduced since the proposed fast affine ME algorithm which will be introduced later on can converge faster under the fourparameter affine motion model. Last but not least, the fourparameter affine motion model can lead to better RD performance for most natural sequences due to the bits savings of header information.
IiiA2 The representation of the fourparameter affine motion model
According to (5), there are four unknown model parameters. Instead of these four parameters, we can also use two MVs to equivalently represent the model because using MVs is more consistent with existing video coding framework. Those two MVs can be chosen at any locations of the current block for representing the motion model. In this paper, we choose the MVs at the top left and top right locations of the current block, because these two locations are adjacent to the previously reconstructed blocks, and the corresponding MVs can be more accurately predicted. In a typical block as shown in Fig. 2, if we denote the MV of top left pixel as and the MV of top right pixel as , the four unknown model parameters , , , and can be solved as follows according to (5).
(6) 
Then (5) can be expressed as a linear combination of and ,
(7) 
where , , , and are equal to , , , and , respectively. and are the horizontal and vertical parts of . It should be noted that , , , and are all related to the coordinate of the current pixel. Eq. (7) can also be written in a vector form,
(8) 
where ,
(9) 
(10) 
Eq. (8) shows that and control the motions of all the pixels in a block. If we know the motions of all the pixels in the block, then the MC process can be performed and the corresponding prediction block can be obtained. Therefore, the key problem becomes how to determine and . In this paper, the precisions of both and are set as pixel to get a good tradeoff between the affine motion model accuracy and overhead bits. This is also consistent with HEVC.
IiiB Affine motion estimation
There are usually two methods to determine the translational MV in a typical encoder of HEVC (HM or x265): AMVP mode combined with a fast ME algorithm and merge mode. The AMVP mode constructs an MVP candidate list for the translational MV and the ME process is used to get the optimal MV for MC. The merge mode constructs a merge candidate list and reuses the motion information of the neighboring blocks. Analogously, we also design two methods in this paper to determine the affine MVs: advanced affine motion vector prediction (AAMVP) mode combined with a fast affine ME method and affine model merge (AMM) mode.
IiiB1 Aamvp
Similar to the AMVP mode, the AAMVP mode tries to obtain a candidate list of MV tuples to predict . The construction of AAMVP candidate list is performed in three steps. Firstly, we find the available MVP candidates for , , and (the MV of the bottom left corner) separately. As shown in Fig. 3, the MVs of neighboring blocks A, B, and C are used as the candidates for the MVP of , the MVs of neighboring blocks D and E are used for the MVP of , and the MVs of neighboring blocks F and G are used for the MVP of . We will use the derivation of the candidates for (the MVP of ) as an example to explain. The availability of the MVs of blocks A, B, and C will firstly be checked (for example, intra mode means unavailable). If available, we will then check whether the MVs of blocks A, B, and C are pointing to the same reference frame as the given one. If yes, the candidates for are found. If not, scaling operations are applied to make the MVs of blocks A, B, and C point to the same reference frame as the given one so as to obtain the candidates for . The derivation processes of and (the MVP of and ) are similar to that of .
Secondly, a candidate list of MV tuples is constructed. The and are combined to get a candidate list. There are two constraints for and . On one hand, the and should not be equal since the equal and means translational motion model. On the other hand, the differences between and in both horizontal and vertical directions should not be larger than a predefined threshold. Too large differences mean and
are probably from different objects, which makes the combination of
and in a single motion model unreasonable. The threshold is set as half of the block size in our implementation. Since we may find multiple candidates through the above steps, they should be put into the candidate list in a specified order. Note that, we have the following relationships between , , and (refer to Fig. 3 and Eq. (5)).(11) 
These relationships can be easily converted to the following constraints for , , and .
(12) 
Since the better MVP tuple will be the one more approximated to the MV tuple, we calculate a criterion named as follows.
(13) 
The smaller value means that the combination of and is more probable to form a real affine motion model and therefore, it should be put in the relatively earlier position in the candidate list.
Thirdly, if the number of MV tuples in the candidate list is less than the maximum number of candidates, each component of the MV tuple set as the translational motion is added to the candidate list to guarantee parsing robustness [29]. The maximum number of MV tuples is set as to be consistent with HEVC AMVP mode. After the above steps, we will have a candidate list of MV tuples with two candidates.
IiiB2 Fast affine ME
After the derivation of MVP tuple candidate lists, the affine ME needs to be performed to find the optimal parameters, i.e. two MVs for a block. A straightforward method is to search all the combinations of and within a predefined search range . However, such a method will lead to the complexity of since two MVs are needed to be jointly determined. Huang et al. [10] provide a simplified fast ME method to optimize one MV out of the two MVs iteratively. In this case, the complexity will be reduced to as the two MVs are optimized independently. However, optimizing one MV out of the two MVs iteratively may be unable to achieve the optimal RD performance. Moreover, since the affine MC is rather complex, the above mentioned ME algorithms are unable to achieve acceptable encoding complexity. Therefore, we propose a gradientbased fast affine ME algorithm which can solve the two MVs simultaneously at each iteration and converge to the optimal combination quickly. The encoding complexity will be determined by the iteration times in the proposed algorithm. And according to our empirical study, and times of iteration will be enough for the unidirectional and bidirectional prediction, respectively. Therefore, the proposed algorithm can reduce the encoder complexity significantly compared with previously studied affine ME algorithms.
The essence of the fast affine ME algorithm is to adjust and to minimize the mean square error (MSE) between the current block and the prediction block. The start search point of affine ME is the best MV tuple among MV tuples in the AAMVP candidate list and the MV tuple with each component equal to the best translational motion. The MSE between the current block and the prediction block can be expressed as
(14) 
where is a collection of all the pixels in the current block, and is the position of the current pixel in the current picture. is the current picture. is the reference picture. is the MV of position .
Define that at the iteration, the MV of position is . Assume that the MVs in the corner positions will change by to obtain the minimum MSE between the current block and the prediction block in the next iteration, then according to (8), the change of MVs for all the pixels in the block can be expressed as
(15) 
Then can be calculated through
(16) 
where is the corresponding position of in the reference block in the iteration. Using the Taylor’s expansion and ignoring the highorder terms, we have
(17) 
As mentioned above, the optimization target is to select the best by minimizing the MSE,
(18) 
Combining (17) and (18), we will have
(19) 
where is equal to . Formula (19) is actually an unconstrained optimization problem. By setting to zero the gradients with respect to , we can obtain
(20) 
where represents the column of matrix . Formula (20) is actually a system of linear equations. can be calculated after the iteration by subtracting the prediction block from the original block. Both and are known values according to (8). is the gradient value at pixel in the reference picture, which can be estimated using the Sobel operator as shown in Eq. (21).
(21) 
Therefore, for each iteration, just a simple system of linear equations needs to be solved to get the . If all components of are after the th iteration, the will be used to get the prediction block. Different from the traditional fast ME algorithms which can only find the best MV one by one, both and can be found out simultaneously through the proposed gradientbased fast affine ME algorithm. Therefore, the proposed algorithm can simultaneously guarantee the RD performance and reduce the encoder complexity significantly.
After the two affine MVs are determined, the affine MVP tuple in the affine MVP tuple list which will lead to smaller affine MVD will be used as the final affine MVP tuple and the two corresponding affine MVDs will be encoded in a similar way as the MVD for the translational motion model. Such a scheme will lead to about two times of bits cost per prediction unit (PU) since two MVDs are transmitted per PU. However, since the affine motion model can improve the prediction accuracy, the number of PUs will reduce significantly due to the use of large blocks, which will lead to less number of MVDs and fewer bits for header information. Besides, the residue bits will also decrease obviously due to the improved prediction precision brought by the affine motion model.
IiiB3 Amm
Different from the existing affine merge mode which tries to regenerate an affine motion model according to the neighboring motion information [17, 20], the AMM mode fully reuses the affine motion model of the neighboring blocks that also use affine mode (including AAMVP and AMM mode). It should be emphasized that the AMM mode is used only when at least one of the neighboring blocks uses affine mode. Fig. 4 gives a typical example showing the difference between the AMM and existing affine merge mode. In Fig. 4, the two green squares represent two neighboring CTUs. In this case, the two CTUs are within the same object that is rotating. Thus, it implies that the two CTUs probably share the same affine motion model parameter . The situation is similar for zooming motion and the zooming factor is the same for neighboring CTUs. Therefore, the reuse of the affine motion model of the neighboring blocks means we can reuse the parameters and in (5) since the zooming factor and rotation angle are the same for the neighboring blocks within one object. Note however that the parameters c and f may be different for neighboring blocks. This is indeed the reuse of model parameters, rather than using the MVs of neighboring blocks, as previous work did [20]. In the previous work, the regenerated model from the neighboring red and green blocks may lead to inaccurate model parameters.
To reuse the affine motion model of the neighboring blocks, we should firstly traverse the neighboring blocks to find the blocks using affine mode. The search order of the AMM candidates is A, B, C, D, and E as shown in Fig. 5, which is the same as the merge mode for translational motion model in HEVC. If A, B, C, D, or E uses affine motion model, the candidate is added to the AMM candidate lists. If no neighboring blocks use affine motion model, the AMM mode will be skipped for the current block. The number of AMM candidates is set as 1 to reduce the header bits for AMM index.
Then we will use the neighboring affine motion parameters to derive the affine motion parameters of the current block. As shown in Fig. 5, since the top left pixel with motion and top right pixel with motion determine the affine motion parameters of the current block, we will deduce the and using the rule of the same and in neighboring blocks. In the following, block A will be used as an example to explain the detailed deduction process. Firstly, we will find the PU containing the block A and obtain the motion information of the PU: the top left pixel with motion , the top right pixel with motion , the bottom left pixel with motion . It should be noted that if block A uses affine mode, it means that , , and are with the same inter direction (forward, backward, or bidirection) and reference index. Then we can calculate the of the current block according to the relative position of the current position with the neighboring PU,
(22) 
Then the of the current block can be calculated using the rule of the same and in the neighboring blocks,
(23) 
The and can be calculated in a similar way if the block B, C, D, or E uses affine mode. The AMM mode with residue and AMM skip mode without residue are both supported in our scheme.
IiiC Fast coding tools for Affine MC
The complexity of affine MC mainly comes from two aspects: the times of interpolation and the complexity of each interpolation. We have mainly designed two coding tools focusing on these two aspects to speed up the affine MC process. The first one is to design a onestep subpixel interpolation filter to decrease the complexity of each interpolation. The second one is to use the affine interpolationprecisionbased adaptive block size MC instead of pixelbased MC to decrease the times of interpolation.
IiiC1 Onestep subpixel interpolation filter
To obtain the affine prediction block with noninteger MVs, for the Luma component, the traditional twostep interpolation filter [10] will first interpolate the pixel accuracy using Discrete Cosine Transform based Interpolation Filter (DCTIF), which is the interpolation filter for translational MC in HEVC. Then the bilinear interpolation will be performed if the MV is beyond pixel accuracy. The situation is similar for Chroma component, the pixel accuracy is firstly interpolated using DCTIF and then the bilinear interpolation is applied for higher pixel accuracy.
The traditional twostep interpolation method mainly has three shortcomings. Firstly, the computational complexity of the traditional method is much higher compared with the translational MC. To interpolate a pixel higher than pixel accuracy, up to four DCTIF and one bilinear interpolation operations should be performed, which will bring quite significant complexity burdens to both the encoder and decoder. Secondly, the bilinear interpolation filter is unable to achieve a satisfactory RD performance for the fractional interpolation [30]. Thirdly, the arbitrary MV precision is unfriendly to the hardware implementation.
To overcome the disadvantages brought by the traditional interpolation filter, we will first determine the MV limitation precision to prevent the unfriendly arbitrary MV precision. The MV limitation precisions of Luma and Chroma components are set as to obtain a balance between RD performance and hardware implementation. According to our empirical study, higher interpolation precision beyond brings out only little improvement in MC accuracy. Besides, a onestep subpixel interpolation filter designed based on the principle of DCTIF is used to interpolate the pixel precision. Since the DCTIF outperforms the bilinear interpolation filter in the aspect of interpolating the fractional pixels [30], the proposed onestep subpixel interpolation filter can achieve better RD performance compared with the traditional twostep interpolation filter. This will also be verified by the experimental results shown in the next section. Moreover, the onestep subpixel interpolation filter can obtain the prediction pixel using only one DCTIF operation for all the pixel precisions and thus can reduce the MC complexity significantly. To unify with the translational interpolation filter in HEVC, the taps of interpolation filter for Luma and Chroma components are set as and , respectively. Due to the limited space, the interpolation filter coefficients for Luma and Chroma components are not shown in this paper. More detailed interpolation coefficients can be found in [31].
IiiC2 Affine interpolationprecisionbased MC
To further reduce the MC complexity, we try to decrease the MV resolution from pixel level to block level to decrease the interpolation times. As shown in Fig. 6 (a), if the size of the MC unit is a pixel, which allows different pixels with various MVs, to interpolate a pixel shown as the small red square, fractional pixels ( horizontal interpolations to obtain the triangles and vertical to obtain the red square) need to be calculated. Therefore, to interpolate a block, we need to perform totally times of interpolation. However, as shown in Fig. 6 (b), if the size of the MC unit is a block, only totally fractional pixels ( horizontal interpolations to obtain the triangles, vertical interpolations to obtain the red squares) need to be calculated. The reduction of the interpolation times mainly comes from the reduction of repetitive interpolation operations for the neighboring pixels with the same MV. It should be noted that the difference between blockbased MC and pixelbased MC will become much more significant if the MC block size becomes larger.
According to the above analysis, we know that decreasing the MV resolution from the pixel level to the block level can reduce the MC complexity significantly. However, it may also lead to some RD performance losses on highorder motion models. To deal with such a problem, an affine interpolationprecisionbased adaptive block size MC scheme is proposed to adapt to the various video characteristics. The basic concept of the affine interpolationprecisionbased adaptive block size MC is that the affine MV precision of the current block will not be lower than a predefined precision. The size of each MC unit is determined by
(24) 
where and are the width and height of the current PU. All the pixels in a PU share the same affine motion parameters. is the relatively larger MV difference between the horizontal and vertical directions for the top left and top right positions. is the corresponding MV difference of the top left and bottom left positions. determines the minimum precision of the affine interpolation, and it is set as in our experiment. It can be seen from (24) that the larger or is, the smaller the size of the MC unit will be to guarantee the affine interpolation precision. The minimum size of an affine MC unit is set as .
Iv Experimental Results
Iva Simulation setup
To evaluate the performance of the proposed fourparameter affine MC framework, the proposed algorithm is implemented in the HEVC reference software HM16.7 [32]. The low delay (LD) main profile and random access (RA) main profile configurations specified in [33] are used as the test conditions. The quantization parameters (QP) tested are , , , and following the HEVC common test condition. Bjontegaard Deltarate (BDrate) [34] is employed in our experiments for fair RD performance comparison.
As the proposed algorithm is designed to better characterize the combination of rotation, zooming, and translation, some sequences with rich rotation or zooming motions are selected to verify the performance of the proposed algorithm. To be more specific, some segments with rich rotation or zooming motions are extracted from the full sequences to better explain the benefits brought by the proposed algorithm. These segments will be called as affine test sequences in the following parts. The detailed characteristics of the affine test sequences are shown in Table I. From Table I, we can see that the affine test sequences include various spatial/temporal resolutions and different characteristics.
Sequence  Picture  Video  Video  Frame 

name  order count  characteristic  resolution  rate 
Tractor(TR)  591690  zooming  1920x1080  25 
Shields(SH)  415514  zooming  1920x1080  50 
Jets(JE)  099  zooming  1280x720  25 
Cactus(CA)  099  rotation  1920x1080  50 
BlueSky(BS)  099  rotation  1920x1080  25 
Station(ST)  099  zooming  1920x1080  25 
SpinCalendar(SC)  099  rotation  1280x720  50 
CatRobot(CR)  099  rotation  3840x2160  60 
RollerCoaster(RC)  099  rotation  4096x2160  60 
IvB Performance
In this subsection, we will first show the performance of the entire fourparameter affine MC framework for the affine test sequences. The entire framework contains all the algorithms introduced in this paper including AAMVP combined with the fast ME algorithm, AMM, and the two fast affine MC tools. Then the performance of the proposed framework will be carefully investigated step by step through the experimental results on affine test sequences in the following three aspects: AAMVP combined with the fast ME algorithm, AMM and AAMVP combined with the fast ME algorithm, and the two fast affine MC coding tools. Both the RD performance and encoder/decoder complexity are shown in details. Finally, we will present some experimental results on the sequences defined in the HEVC common test conditions.
IvB1 The performance of the entire framework on the affine test sequences compared with HM16.7 anchor
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –22.5%  –20.1%  –20.1%  –33.4%  –29.4%  –28.8% 
SH  –12.5%  –9.5%  –9.8%  –19.2%  –15.1%  –17.6% 
JE  –6.5%  –6.1%  –5.7%  –24.5%  –24.5%  –24.2% 
CA  –6.0%  –5.2%  –4.8%  –7.3%  –6.5%  –6.6% 
BS  –6.7%  –7.3%  –6.5%  –10.2%  –10.6%  –8.6% 
ST  –21.7%  –19.3%  –19.4%  –35.8%  –31.7%  –32.7% 
SC  –14.6%  –15.1%  –13.8%  –33.0%  –34.8%  –33.4% 
CR  –4.9%  –4.8%  –4.7%  –5.4%  –5.2%  –4.5% 
RC  –4.2%  –4.5%  –4.0%  –4.6%  –4.9%  –4.7% 
Avg.  –11.1%  –10.2%  –9.9%  –19.3%  –18.1%  –17.9% 
EncT  121%  131%  
DecT  112%  123% 
Both the RD performance and encoding/decoding time increase of the entire affine MC framework for affine test sequences are illustrated in Table II. From Table II, we can see that the proposed framework can achieve averagely and BDrate reductions on Y component for the affine test sequences in RA and LD cases, respectively. The experimental results obviously demonstrate that the proposed fourparameter affine motion model can well represent the combination of rotation, zooming, and translation.
Besides the RD performance, the encoding/decoding complexity is another important factor to be measured. From Table II, we can see that the encoding time is about to , and the decoding time is about to compared with the HM16.7 anchor. The experimental results obviously show that the proposed affine MC framework will not increase the encoding and decoding complexity significantly.
IvB2 AAMVP combined with the fast Affine ME
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –22.5%  –19.3%  –19.1%  –31.0%  –27.0%  –26.1% 
SH  –10.4%  –8.8%  –8.9%  –14.2%  –12.1%  –13.3% 
JE  –4.6%  –4.8%  –4.6%  –18.7%  –19.9%  –18.3% 
CA  –7.2%  –6.2%  –5.4%  –7.8%  –7.4%  –7.4% 
BS  –6.8%  –6.7%  –6.3%  –9.8%  –8.4%  –8.6% 
ST  –19.8%  –16.8%  –16.7%  –30.8%  –26.9%  –27.4% 
SC  –14.0%  –13.6%  –12.5%  –29.5%  –32.3%  –30.6% 
CR  –6.0%  –4.9%  –4.5%  –5.6%  –3.9%  –3.4% 
RC  –3.2%  –3.1%  –2.7%  –3.5%  –3.3%  –3.5% 
Avg.  –10.5%  –9.4%  –9.0%  –16.8%  –15.7%  –15.4% 
EncT  232%  296%  
DecT  250%  409% 
The overall RD performance and encoding/decoding time increase of only enabling the AAMVP combined with the fast affine ME for affine test sequences compared with HM16.7 anchor are illustrated in Table III. From Table III, we can see that the proposed AAMVP combined with the fast affine ME algorithm can provide averagely and bitrate reductions on Y component in RA and LD cases, respectively. From the results in Table III and Table II, we can obviously see that the AAMVP mode contributes most of the BDrate reductions provided by the fourparameter affine MC framework. For the encoding/decoding time, from Table III, we can see that the encoding time is about to compared with the HM16.7 anchor. Besides, the decoding time is about to accordingly. Both the encoding and decoding burdens are quite high since the fast affine MC coding tools are not used in the test.
IvB3 AMM and AAMVP combined with the fast ME algorithm
The AMM mode always attempts to reuse the affine motion model of the neighboring blocks using affine mode. Therefore, the AMM mode cannot be used without AAMVP. In this subsection, we will see both the RD performance improvement and encoding/decoding time change brought by the AMM mode on the premise of AAMVP.
The average bitrate savings and encoding/decoding time increase of the AMM and AAMVP combined with the fast affine ME algorithm for affine test sequences compared with HM16.7 anchor are shown in Table IV. Through the comparison between Table IV and Table III, we can see that the proposed AMM mode can further achieve about and RD performance improvements in average on the premise of AAMVP mode in RA and LD cases, respectively. The experimental results demonstrate that the proposed AMM mode is beneficial to the overall RD performance. As for the encoding/decoding complexity, the encoding time of the AMM combined with AAMVP only increases a little compared with the AAMVP mode since the AMM mode without complex affine ME operations is quite simple. However, the decoding time increases very obviously. The reason is that the number of blocks choosing affine mode increases significantly, and then the computations of affine MC are multiplied. The average decoding time for the affine test sequences can be as much as compared with the HEVC anchor in LD case. Note that in this test, the fast affine MC coding tools are still not used.
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –24.0%  –21.0%  –21.0%  –34.7%  –31.4%  –30.2% 
SH  –13.5%  –11.1%  11.1%  –19.7%  –17.7%  –18.9% 
JE  –6.1%  –5.4%  –5.2%  –23.5%  –24.4%  –23.2% 
CA  –8.0%  –6.7%  –6.1%  –8.7%  –8.3%  –8.3% 
BS  –8.1%  –7.8%  –7.2%  –11.1%  –10.1%  –9.8% 
ST  –22.7%  –19.6%  –19.7%  –37.1%  –32.8%  –33.6% 
SC  –15.8%  –16.1%  –14.4%  –32.5%  –35.5%  –33.9% 
CR  –6.7%  –5.9%  –5.3%  –6.0%  –4.8%  –4.1% 
RC  –4.7%  –4.7%  –4.3%  –4.7%  –4.9%  –4.6% 
Avg.  –12.2%  –10.9%  –10.5%  –19.8%  –18.9%  –18.5% 
EncT  236%  301%  
DecT  326%  523% 
IvB4 The performance of onestep interpolation filter
In this subsubsection, the performances of the proposed two fast affine MC coding tools are presented. The overall RD performance and encoding/decoding time change of the onestep subpixel interpolation filter for affine test sequences compared with HM16.7 anchor are shown in Table V. Through the comparison between Table IV and Table V, we can see that the DCTIF based onestep subpixel interpolation filter can further bring about and bitrate savings averagely on Y component for affine test sequences in RA and LD cases compared with the situation using bilinear interpolation filter. The benefits mainly come from the fact that the DCTIF can bring better interpolation results than the bilinear interpolation filter when the MV precision is higher than pixel. Besides, since the onestep subpixel interpolation filter can reduce the interpolation times dramatically for those blocks with affine MVs higher than pixel precision, the encoding and decoding complexities are both reduced significantly. Especially, for the affine test sequences in LD case, the decoding time reduces from to .
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –24.0%  –21.1%  –21.0%  –35.0%  –31.4%  –30.3% 
SH  –13.6%  –10.7%  –10.5%  –20.1%  –16.9%  –19.0% 
JE  –6.8%  –6.2%  –5.6%  –25.4%  –25.8%  –24.2% 
CA  –8.4%  –7.1%  –6.7%  –9.2%  –8.4%  –8.4% 
BS  –8.3%  –8.7%  –7.7%  –12.0%  –11.6%  –10.1% 
ST  –23.0%  –19.6%  –20.0%  –37.3%  –32.1%  –33.9% 
SC  –17.0%  –16.6%  –15.7%  –36.3%  –37.5%  –37.1% 
CR  –6.8%  –6.0%  –5.7%  –6.2%  –5.4%  –4.6% 
RC  –4.7%  –4.6%  –4.2%  –4.8%  –5.0%  –4.9% 
Avg.  –12.5%  –11.2%  –10.8%  –20.7%  –19.4%  –19.2% 
EncT  136%  153%  
DecT  135%  170% 
IvB5 The performance of adaptive block size MC
The average performance and encoding/decoding time change of the affine interpolationprecisionbased adaptive block size MC for affine test sequences compared with HM16.7 anchor are shown in Table VI. Through the comparison between Table IV and Table VI, we can see that the BDrate increases by about and in RA and LD cases in average for affine test sequences compared with the situation without any fast MC algorithms. The losses mainly come from the fact that the minimum size of the MC unit is set as to reduce the computational burden for both the hardware and software implementation. Although the adaptive block size MC will incur a few performance losses, it can bring quite a significant encoding/decoding complexity reduction. As an example, the decoding time decreases from to for affine test sequences in LD case.
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –22.4%  –20.1%  –19.8%  –33.0%  –28.5%  –28.3% 
SH  –12.5%  –10.3%  –10.3%  –18.9%  –16.5%  –18.6% 
JE  –6.0%  –5.3%  –5.3%  –22.4%  –23.7%  –22.5% 
CA  –5.7%  –5.0%  –4.5%  –6.7%  –6.6%  –6.2% 
BS  –6.5%  –6.4%  –5.9%  –9.1%  –8.5%  –7.5% 
ST  –21.4%  –19.2%  –19.0%  –35.6%  –32.2%  –31.9% 
SC  –13.2%  –13.9%  –12.6%  –29.2%  –32.3%  –30.9% 
CR  –5.0%  –4.8%  –4.5%  –5.3%  –4.7%  –4.2% 
RC  –4.3%  –4.7%  –4.0%  –4.6%  –4.5%  –4.3% 
Avg.  –10.8%  –10.0%  –9.5%  –18.3%  –17.5%  –17.1% 
EncT  138%  156%  
DecT  154%  199% 
IvB6 The performance of two fast affine MC coding tools together
The overall RD performance and encoding/decoding complexity of the combined algorithm for the affine test sequences compared with HM16.7 anchor have already been shown in Table II in advance. Compared with the situation without any fast MC algorithms as shown in Table IV, the fast affine MC algorithms totally just suffer about and RD performance losses for the affine test sequences in RA and LD cases, accordingly. Although there are a few performance losses, the encoding/decoding complexity decrease is quite amazing. The encoding time is only about to increase compared with the HM16.7 anchor without affine mode. The decoding time increases about to for affine test sequences. The experimental results obviously demonstrate that the combination of the two fast affine MC coding tools can lead to a better tradeoff between the RD performance and encoding/decoding complexity.
IvB7 RD curve
Some example RD curves are shown in Fig. 7. These RD curves also obviously demonstrate that the proposed algorithm can achieve much better RD performance improvement compared with the HEVC anchor. It can be obviously seen from Fig. 7 that the performance improvement mainly comes from the bitrate reduction instead of YPSNR improvement. Besides, we can also see that the entire framework presents ignorable RD performance losses compared with the “AMM and AAMVP” (without fast MC coding tools) in the figure.
IvB8 The performance of the entire framework on the affine test sequences compared with the previous work
Affine  RA  LD  

Class  Y  U  V  Y  U  V 
TR  –14.6%  –13.4%  –13.1%  –22.5%  –20.1%  –20.1% 
SH  –3.4%  –2.9%  –3.2%  –12.5%  –9.5%  –9.8% 
JE  –2.3%  –2.0%  –1.9%  –6.5%  –6.1%  –5.7% 
CA  –3.5%  –2.9%  –2.8%  –6.0%  –5.2%  –4.8% 
BS  –3.1%  –3.6%  –3.1%  –6.7%  –7.3%  –6.5% 
ST  –11.2%  –11.0%  –10.7%  –21.7%  –19.3%  –19.4% 
SC  –7.3%  –7.3%  –6.2%  –14.6%  –15.1%  –13.8% 
CR  –2.0%  –1.4%  –1.8%  –3.6%  –1.8%  –2.0% 
RC  –1.8%  –1.6%  –1.5%  –1.0%  –1.6%  –1.2% 
Avg.  –5.5%  –5.1%  –5.0%  –6.8%  –5.1%  –4.7% 
EncT  419%  323%  
DecT  507%  951% 
We also compare the entire framework with our previous work [11] to demonstrate the effectiveness of the proposed algorithm in this paper. Table VII gives the detailed experimental results of our previous work on the affine test sequences. From Table VII, we can see that the proposed algorithm can bring averagely and RD performance improvements compared with the HEVC anchor for the affine test sequences. Compared with the results of the entire framework shown in Table II, we can see that the proposed algorithm can bring over double BDratesavings compared with the previous work. The significant bitrate savings of the proposed method mainly comes from the following five aspects.

The reduction of the parameters from 6 to 4 can effectively reduce the header information so as to improve the RD performance.

In the proposed work in this paper, we have searched over all the reference frames in list0 and list1, and the combination of them to obtain better prediction block. However, in the previous work, we have not focused on developing any fast affine MC algorithms. Therefore, to reduce encoding complexity, we only search the inter direction and the reference frame, which are the same as those of the top left block.

The AAMVP can achieve better MVP tuple to accurately predict the affine MV.

The AMM mode can save lots of header information.

The DCTIF based one step interpolation filter can lead to better coding efficiency compared with the bilinear interpolation filter used in the previous work.
Besides, from the encoding/decoding complexity point of view, the proposed algorithm also achieves much lower encoding and decoding complexity compared with the previous work.
IvB9 Performance of the HEVC common test condition sequences
General  RA  LD  

Class  Y  U  V  Y  U  V 
Class A  –0.5%  –0.4%  –0.4%  –  –  – 
Class B  –1.4%  –1.2%  –1.1%  –1.5%  –1.5%  –1.3% 
Class C  –0.7%  –0.7%  –0.9%  –1.0%  –1.0%  –1.3% 
Class D  –1.1%  –1.2%  –1.3%  –2.0%  –2.2%  –2.8% 
Class E  –  –  –  –1.8%  –1.6%  –2.0% 
Class F  –1.3%  –1.5%  –1.5%  –1.5%  –1.6%  –1.9% 
Avg.  –1.0%  –0.9%  –0.9%  –1.5%  –1.6%  –1.8% 
EncT  118%  128%  
DecT  103%  105% 
We also present some experimental results on the HEVC common test condition sequences as shown in Table VIII. From Table VIII, we can obviously see that the proposed framework can bring averagely and RD performance improvements in RA and LD cases compared with the HM16.7 anchor for the HEVC common test condition sequences. Besides, the increase of the complexity of the HEVC common test condition sequences, especially the decoding time, is only marginal. Therefore, the proposed technique is promising to be integrated into future video coding standards.
IvC Performance analysis
In this subsection, the benefits brought by the proposed affine MC framework will be analyzed carefully from the change of the following two factors before and after the use of the affine motion model: the coding unit (CU) partition size, the number of blocks using affine mode.
Fig. 8 and Fig. 9 show the CU partition of an inter frame of a typical affine sequence for HEVC anchor and the proposed fourparameter affine motion model framework, respectively. In both figures, a red square represents a CU and picture order count (POC) means the frame number in display order. From these two figures, we can obviously see that the CU partitions are quite small for most of the blocks for HEVC anchor while the CU partitions become quite large when the affine motion model is applied. The reason is that the zooming motion in the sequence can be well represented by the proposed affine motion model and thus large CU partitions can be used. However, for the HEVC anchor with only translational motion model, the encoder has to split a block into smaller ones so that for each smaller block the motion is approximately translational. Therefore, using our proposed affine motion model can enable the use of large blocks and thus reduce the overhead bits on block partitions significantly.
Fig. 10 and Fig. 11 show the situations of the blocks using affine motion model in the sequences with global and local affine motion, respectively. In both figures, the red squares represent the CUs using affine mode. In the sequence with global affine motion, almost all the blocks choose the affine mode, with only two kinds of blocks being exceptions. The first kind is the blocks in the border of a frame for which a good prediction is impossible to be obtained. The second kind is the very smooth blocks such as blue sky for which using translational motion model can also lead to a good prediction. For the sequence with local affine motion, most of the blocks with rotation motion choose the affine mode while the other blocks with translational motion choose the translation motion model. The encoder can determine a suitable motion model for each block through RDO. The experimental results obviously demonstrate that the proposed affine MC framework combined with the traditional translational MC framework can well represent various video contents with global or local complex motions.
Table IX gives the average affine mode percentages for all the frames for the affine test sequences. It can be seen from Table IX that the affine mode percentages can be quite high for the affine test sequences. Also, we can see from Table IX that the affine mode percentages in RA case are obviously lower than those in LD case. This is mainly due to the fact that the utilization of both the forward and backward reference frames is beneficial for obtaining a better prediction for the blocks with complex motions. Therefore, the benefits brought by the proposed affine motion model in RA case are less than those in LD case.
Affine Class  RA  LD 

Tractor  30.0%  53.7% 
Shields  23.7%  49.7% 
Jets  11.6%  31.3% 
Cactus  8.1%  10.6% 
BlueSky  33.4%  53.1% 
Station  34.9%  65.8% 
SpinCalendar  24.3%  62.4% 
CatRobot  5.9%  9.0% 
RollerCoaster  12.4%  15.2% 
V Conclusion
In this paper, an effective fourparameter affine motion compensation framework is proposed to better characterize the combination of rotation, zooming, and translation. In the framework, a fourparameter affine motion model is firstly proposed and analyzed. Then the four parameters are proposed to be coded in two manners: advanced affine motion vector prediction and affine model merge. Especially, different from the traditional merge mode to regenerate a new affine motion model using the neighboring motion information, the affine model merge mode reuses the affine motion model of the neighboring blocks using affine mode. Moreover, two fast motion compensation tools including onestep subpixel interpolation filter and affine interpolationprecisionbased adaptive block size motion compensation are proposed to speed up the affine motion compensation process. The proposed framework is implemented in the reference software of High Efficiency Video Coding (HEVC). The experimental results show that the proposed affine motion compensation framework can achieve much better rate distortion performances compared with the HEVC anchor for the sequences with rich rotation or zooming motions. The experimental results demonstrate the effectiveness of the proposed affine motion compensation framework.
In the current implementation, we only focus on a quite simple fourparameter affine motion model to characterize the combination of rotation, zooming, and translation. However, how to effectively characterize other complex motions using highorder motion models remains an open issue. Besides, the global highorder motion model is sometimes more effective than the local highorder motion model. We will try to integrate the global and local highorder motion models into a whole framework in our future work.
References
 [1] T. Wiegand, G. Sullivan, G. Bjontegaard, and A. Luthra, “Overview of the H.264/AVC video coding standard,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 13, no. 7, pp. 560–576, Jul. 2003.
 [2] G. Sullivan, J. Ohm, W.J. Han, and T. Wiegand, “Overview of the High Efficiency Video Coding (HEVC) Standard,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 22, no. 12, pp. 1649–1668, Dec. 2012.
 [3] I.K. Kim, J. Min, T. Lee, W.J. Han, and J. Park, “Block partitioning structure in the HEVC standard,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 22, no. 12, pp. 1697–1706, Dec. 2012.
 [4] V. E. Seferidis and M. Ghanbari, “General approach to blockmatching motion estimation,” Optical Engineering, vol. 32, no. 7, pp. 1464–1474, 1993.
 [5] T. Wiegand, E. Steinbach, and B. Girod, “Affine multipicture motioncompensated prediction,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 15, no. 2, pp. 197–209, Feb. 2005.
 [6] X. Li, J. R. Jackson, A. K. Katsaggelos, and R. M. Merserau, “Multiple global affine motion model for H.264 video coding with low bit rate,” Proceedings of SPIE  The International Society for Optical Engineering, vol. 5685, pp. 185–194, 2005.
 [7] Y. Nakaya and H. Harashima, “Motion compensation based on spatial transformations,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 4, no. 3, pp. 339–356, 366–7, Jun. 1994.
 [8] C.L. Huang and C.Y. Hsu, “A new motion compensation method for image sequence coding using hierarchical grid interpolation,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 4, no. 1, pp. 42–52, Feb. 1994.
 [9] H.K. Cheung and W.C. Siu, “Local affine motion prediction for H.264 without extra overhead,” in Circuits and Systems (ISCAS), Proceedings of 2010 IEEE International Symposium on, May 2010, pp. 1555–1558.
 [10] H. Huang, J. Woods, Y. Zhao, and H. Bai, “Controlpoint representation and differential coding affinemotion compensation,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 23, no. 10, pp. 1651–1660, Oct. 2013.
 [11] L. Li, H. Li, Z. Lv, and H. Yang, “An affine motion compensation framework for high efficiency video coding,” in 2015 IEEE International Symposium on Circuits and Systems (ISCAS), May 2015, pp. 525–528.
 [12] H. Yu, Z. Lin, and F. Teo, “An efficient coding scheme based on image alignment for H.264/AVC,” in Circuits and Systems, 2009. ISCAS 2009. IEEE International Symposium on, May 2009, pp. 629–632.
 [13] C. Toklu, A. Erdem, M. Sezan, and A. Tekalp, “Tracking motion and intensity variations using hierarchical 2D mesh modeling for synthetic object transfiguration,” Graphical Models and Image Processing, vol. 58, no. 6, pp. 553–573, 1996.
 [14] G. AlRegib, Y. Altunbasak, and R. Mersereau, “Hierarchical motion estimation with contentbased meshes,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 13, no. 10, pp. 1000–1005, Oct. 2003.
 [15] R. Kordasiewicz, M. Gallant, and S. Shirani, “Affine motion prediction based on translational motion vectors,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 17, no. 10, pp. 1388–1394, Oct. 2007.
 [16] M. Narroschke and R. Swoboda, “Extending hevc by an affine motion model,” in 2013 Picture Coding Symposium (PCS), Dec 2013, pp. 321–324.
 [17] H. Huang, J. Woods, Y. Zhao, and H. Bai, “Affine skip and direct modes for efficient video coding,” in Visual Communications and Image Processing (VCIP), 2012 IEEE, Nov. 2012, pp. 1–6.
 [18] C. Heithausen and J. H. Vorwerk, “Motion compensation with higher order motion models for HEVC,” in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), April 2015, pp. 1438–1442.
 [19] P. Helle, S. Oudin, B. Bross, D. Marpe, M. Bici, K. Ugur, J. Jung, G. Clare, and T. Wiegand, “Block merging for quadtreebased partitioning in HEVC,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 22, no. 12, pp. 1720–1731, Dec. 2012.
 [20] H. Chen, F. Liang, and S. Lin, “Affine skip and merge modes for video coding,” in Multimedia Signal Processing (MMSP), 2015 IEEE 17th International Workshop on, Oct. 2015, pp. 1–5.
 [21] H. Yuan, Y. Chang, Z. Lu, and Y. Ma, “Model based motion vector predictor for zoom motion,” Signal Processing Letters, IEEE, vol. 17, no. 9, pp. 787–790, Sept. 2010.
 [22] H. Yuan, J. Liu, J. Sun, H. Liu, and Y. Li, “Affine model based motion compensation prediction for zoom,” Multimedia, IEEE Transactions on, vol. 14, no. 4, pp. 1370–1375, Aug. 2012.
 [23] L.M. Po, K.M. Wong, K.W. Cheung, and K.H. Ng, “Subsampled blockmatching for zoom motion compensated prediction,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 20, no. 11, pp. 1625–1637, Nov. 2010.
 [24] H.S. Kim, J.H. Lee, C.K. Kim, and B.G. Kim, “Zoom motion estimation using blockbased fast local area scaling,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 22, no. 9, pp. 1280–1291, Sept. 2012.
 [25] S. Zhu and K.K. Ma, “A new diamond search algorithm for fast blockmatching motion estimation,” Image Processing, IEEE Transactions on, vol. 9, no. 2, pp. 287–290, Feb. 2000.
 [26] C. Zhu, X. Lin, and L.P. Chau, “Hexagonbased search pattern for fast block motion estimation,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 12, no. 5, pp. 349–355, May 2002.
 [27] A. M. Tourapis, “Enhanced predictive zonal search for single and multiple frame motion estimation,” pp. 1069–1079, 2002.
 [28] N. Purnachand, L. Alves, and A. Navarro, “Improvements to TZ search motion estimation algorithm for multiview video coding,” in Systems, Signals and Image Processing (IWSSIP), 2012 19th International Conference on, Apr. 2012, pp. 388–391.
 [29] B. Li, J. Xu, and H. Li, “Parsing robustness in High Efficiency Video Coding  analysis and improvement,” in Visual Communications and Image Processing (VCIP), 2011 IEEE, Nov. 2011, pp. 1–4.
 [30] H. Lv, R. Wang, X. Xie, H. Jia, and W. Gao, “A comparison of fractionalpel interpolation filters in HEVC and H.264/AVC,” in Visual Communications and Image Processing (VCIP), 2012 IEEE, Nov. 2012, pp. 1–6.
 [31] S. Lin, H. Chen, H. Zhang, S. Maxim, H. Yang, and J. Zhou, “Affine transform prediction for next generation video coding,” ITUT SG16 Doc.COM16C1016, Oct. 2015.
 [32] HM, HEVC test Model. [Online]. Available: {https://hevc.hhi.fraunhofer.de/svn/}
 [33] F. Bosson, “Common test conditions and software reference configurations,” Document JCTVCL1100, Geneva, CH, Jan. 2013.
 [34] G. Bjontegaard, “Calculation of average PSNR differences between RDcurves,” Document VCEGM33, Austin, Texas, USA, Apr. 2001.
Comments
There are no comments yet.