Next Article in Journal
Prediction of Air Quality Combining Wavelet Transform, DCCA Correlation Analysis and LSTM Model
Next Article in Special Issue
Lightweight Tennis Ball Detection Algorithm Based on Robomaster EP
Previous Article in Journal
Evaluation of Body Composition Changes by Bioelectrical Impedance Vector Analysis in Volleyball Athletes Following Mediterranean Diet Recommendations during Italian Championship: A Pilot Study
Previous Article in Special Issue
Automatic Ship Object Detection Model Based on YOLOv4 with Transformer Mechanism in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Network-Based Reference Block Quality Enhancement for Motion Compensation Prediction

School of Control Science and Engineering, Shandong University, No. 17923 Jingshi Road, Jinan 250061, China
*
Author to whom correspondence should be addressed.
Submission received: 19 December 2022 / Revised: 19 February 2023 / Accepted: 20 February 2023 / Published: 22 February 2023
(This article belongs to the Special Issue Computer Vision and Pattern Recognition Based on Deep Learning)

Abstract

:
Inter prediction is a crucial part of hybrid video coding frameworks, and it is used to eliminate redundancy in adjacent frames and improve coding performance. During inter prediction, motion estimation is used to find the reference block that is most similar to the current block, and the following motion compensation is used to shift the reference block fractionally to obtain the prediction block. The closer the reference block is to the original block, the higher the coding efficiency is. To improve the quality of reference blocks, a quality enhancement network (RBENN) that is dedicated to reference blocks is proposed. The main body of the network consists of 10 residual modules, with two convolution layers for preprocessing and feature extraction. Each residual module consists of two convolutional layers, one ReLU activation, and a shortcut. The network uses the luma reference block as input before motion compensation, and the enhanced reference block is then filtered by the default fractional interpolation. Moreover, the proposed method can be used for both conventional motion compensation and affine motion compensation. Experimental results showed that RBENN could achieve a −1.35% BD rate on average under the low-delay P (LDP) configuration compared with the latest H.266/VVC.

1. Introduction

As video dominates the current media transmission network, the demand for higher quality video is growing day by day, which brings great challenges to video compression algorithms. Video coding tools have been iterating for decades and, furthermore, eliminating redundant information in the video is becoming more and more difficult.
As one of the most important parts in a hybrid video coding framework, inter prediction predicts the correlation between adjacent frames so that it is not necessary to transmit all the information of each coding frame. As shown in Figure 1, when encoding the current coding unit (CU), the reference CU, which is most similar to the current CU, can be found in the reconstructed reference frame. Only the motion vector (MV) and residual information between them need to be transmitted. Due to the inherent spatial discretization of digital video, block translation may not be aligned with integer pixels. Through motion estimation (ME), as shown in Figure 2, the current CU first finds the reference CU and obtains the integer pixel MV ( I m v ). To improve the prediction accuracy, the reference CU is further interpolated to achieve a higher resolution; thus, the image block at the best fractional pixel position could be selected by a fractional MV ( F m v ). It is obvious that the interpolation method plays an important role for fractional motion estimation. The more accurate the predicted block, the smaller the residue, and the higher the coding performance. In the 1980s, researchers adopted simple bilinear and bicubic filters to generate fractional pixels. With the development of signal processing theory, researchers proposed more efficient interpolation filters, such as the Wiener interpolation filter, discrete cosine transform interpolation filter (DCTIF) [1], and the generalized interpolation filter [2]. These filters can perform an overall fractional shift of the image to make the reference block closer to the current block. In addition, for non-translational motion in video, H.266/VVC also uses affine motion compensation, in which a large block is divided into sub-blocks first before the corresponding MV of each sub-block can be calculated by an affine formula to cope with the motion of zooming, rotation, etc. A typical video coding structure of existing video coding standards is shown in Figure 3.
Finally, the optimal MV, the transformed and quantized residual between the current block and the reference block, and some other meta information is entropy coded to generate a bit stream. The decoder finds the corresponding reference block by the MV and adds residuals to reconstruct the current block.
In recent years, deep learning (DL) has made great achievements in many fields with its strong learning ability and robustness. Especially in the field of computer vision, deep convolutional neural networks have played an interesting role in solving image processing problems. Recently, Moving Picture Expert Group has been investigating deep learning-based coding tools to further improve coding efficiency. DL-based methods can be integrated into a specific video coding module, such as intra prediction, inter prediction, loop filter, and post-processing. Specifically, in inter prediction, DL can be used for prediction, quality enhancement, fractional interpolation, and reference frame generation.
In view of the successful application of DL in image quality enhancement, we proposed a quality enhancement method for CU before MC, which was added to the H.266/VVC framework. Specifically, we designed a reference-block-enhanced neural network (RBENN) to enhance the quality of the reference block. Fractional interpolation and other subsequent operations were then used on the quality-enhanced reference block. Compared with the standard H.266/VVC framework, the proposed method can reduce the Bjøntegaard delta rate (BD rate) [3] in the low-delay P (LDP) configuration by 1.35%. The contributions of this paper are as follows.
  • A convolutional neural network with residual structure was used to enhance the quality of reference blocks before MC.
  • The proposed method not only enhanced the reference CU before conventional motion compensation, but it also enhanced the reference CU before affine motion compensation.
  • Experimental results showed that the proposed method achieved a 1.35% BD rate saving compared with the H.266/VVC standard under LDP test conditions.
The remainder of this paper is organized as follows. Section 2 provides a brief review of the related work, Section 3 describes the proposed methodology and network architecture, Section 4 shows experimental results, and the work is concluded in Section 5.

2. Related Work

In inter prediction, the closer the reference CU is to the current CU, the less the residuals are that are required for coding, and, thus, the lower the bit rate. However, due to the discretization of digital video, the current CU and the reference CU cannot be completely aligned. For this reason, existing video coding standards usually adopt a set of fixed sub-pixel filtering coefficients to interpolate the reference CUs [4]. In H.266/High Efficiency Video Coding (HEVC) and H.266/VVC, discrete cosine transform-based half or quarter interpolation filters are adopted. Specifically, to improve the coding performance, H.266/VVC applies one-sixth sub-pixel precision interpolation filters and affine motion compensation.
Although the fractional interpolation filter achieves good performance in ME and MC, it still cannot adapt to complex video content well. Therefore, many researchers have used DL to refine the sub-pixel blocks. In these methods, [5,6,7,8,9,10,11,12,13] are for HEVC/H.265 [14], while [15,16,17,18,19] are for H.266/VVC [20].
Yan et al. [5,6,7] regarded the fractional interpolation filter as an image regression problem and trained neural networks for each fractional interpolation filter. In [5], the author replaced three half-pixel fractional filters with three super-resolution convolutional neural network (SRCNN) [21] models, which resulted in a 0.9% reduction in BD rate under the LDP configuration of H.265/HEVC. In [6], the author used the reference CU as the input, the original CU as the ground truth, and trained 15 networks, whose structures were similar to a variable-filter-size residue learning convolutional neural network (VRCNN) [22], to replace the fractional filters of half- and quarter-pixel fractional shifts. In this method, there are a total of 120 natural networks for uniprediction, biprediction, and four different quantization parameters (QPs). This method saves 3.9%, 2.7%, and 1.3% in BD rates on average for LDP, low-delay B (LDB), and random-access (RA) configurations of H.265/HEVC, respectively. Yan et al. [7] also proposed a reversible training method that is based on the assumption that integer pixels and fractional pixels can be derived from each other. This method can save 4.7% and 3.6% in BD rates under the configurations of LDB and RA for H.265/HEVC, respectively. In [8], a group variational transformation neural network (GVTNN) was proposed to replace the fractional interpolation in H.265/HEVC, and it can achieve a 1.9% BD rate reduction under the LDP configuration. Liu et al. [9] optimized a neural network based on [8] and analyzed the influence of a fuzzy kernel on the task. In [10], Huo et al. deemed that fractional interpolation can obtain more information from the reconstructed area around the current CU. The authors then put the current CU, together with the left and top reconstructed areas, into a larger square area as the network input and the original CU as the ground truth label. This method reduces the BD rate by 1.8% in the LDP configuration of H.265/HEVC. Similarly, Wang et al. [11] reshaped the left and the upper reconstructed regions of the current CU into images with the same size, and then input them into the neural network together. This method can achieve 4.6%, 3.0%, and 2.7% BD rate reductions on average under LDP, LDB, and RA configurations, respectively, compared to H.265/HEVC. Chi et al. [12] trained separate networks for luma and chroma channels. The method in [13] is similar to a loop filter; it uses the optical flow network (PWC-Net) [23] to generate an optical flow graph from the video frames before the reference frames, then uses the optical flow graph and the reference frame in a neural network [24] for quality enhancement. T he reference CUs are then extracted from the quality-enhanced reference frame for MC.
In the study of H.266/VVC, Murn et al. [15,16] designed a three-layer neural network that only contains convolutional layers. The reference CUs, and the original CUs corresponding to different fractional filters, are made into different datasets for network training. All convolution kernels in the network are extracted and compressed into a  13 × 13 matrix to replace the DCTIF for MC. This approach can effectively reduce the complexity of the network; it can finally achieve BD rate savings of 0.77%, 1.27%, and 2.25% on average under RA, LDB, and LDP configurations, respectively, compared to H.266/VVC. Galpin [17] used neural networks to enhance inter biprediction and replaced bidirectional optical flow (BDOF) with networks. Compared with H.266/VVC, it saves 1.4% and 0.58% of the BD rate under RA and LDB configurations, respectively. Jin [18] combined DL with affine motion compensation. This method takes the reference region, the predicted region, and the motion vectors of sub-CUs of affine motion compensation as the inputs of the network, and then outputs the final predicted region. The authors trained 76 networks for different QPs and CU sizes. It finally achieved a −1.77% BD rate compared to H.266/VVC under LDP configuration. In [19], Katayama used LSTM [25] to enhance the reference frame to generate a virtual frame. This method achieves a BD rate reduction of 0.58% compared to H.266/VVC under the LDP configuration.
Most of the above-mentioned methods use a neural network to replace the DCTIF. Unlike these methods, we embedded the RBENN as a module into inter prediction, which could be used together with these methods. To maximize the coding efficiency, we only enhanced the CU blocks whose size was larger than  16 × 16 , and also enhanced the CU blocks before affine motion compensation.

3. Proposed Method

Section 3.1 describes how the RBENN is embedded into the conventional motion compensation and affine motion compensation of H.266/VVC. The composition of the network is then described in Section 3.2. Finally, the dataset and the training method are described in Section 3.3 and Section 3.4.

3.1. An RBENN Integrated in H.266/VVC

In the configuration file of H.266/VVC, CU sizes range from  4 × 4 to  64 × 64 . As the variation of CUs with small sizes is usually not large and the number of pixels is also small, an RBENN only enhances the CUs that are larger than  16 × 16 . We trained three networks, denoted as RBENN 1, RBENN 2, and RBENN 3, to process CU blocks with minimum side lengths of 16, 32, and 64, respectively. Specifically, RBENN1 enhances CUs with sizes of { 16 × 16 32 × 16 16 × 32 16 × 64 64 × 16 }, RBENN2 enhances CUs with sizes of { 32 × 32 64 × 32 32 × 64 }, and RBENN3 enhances CUs with a size of { 64 × 64 }. Figure 4 depicts the system diagram of the encoder and decoder when the neural networks are integrated into the H.266/VVC. On the encoder side, the current CU finds the reference CU based on ME. An RBENN enhances the reference CU and replaces the original reference CU with its output. At the decoder, the current CU finds the reference CU in the reconstructed frame based on the decoded MV. The output of the RBENN is then filtered by a DCTIF to obtain the predicted CU.
Figure 5a shows the process of integrating an RBENN into conventional motion compensation, while Figure 5b is the process of integrating an RBENN into affine motion compensation. In conventional motion compensation, the H.266/VVC standard finds the reference CU from the current CU according to the MV. First, the reference CU is expanded by 4 pixels, then a DCTIF is applied to generate the predicted CU. The residual CU is obtained by subtracting the predicted CU from the original CU. It is then transformed and quantized. Our method was designed to input a padded image patch into the proposed RBENN, and the output image patch is then interpolated and filtered by a DCTIF. In affine motion compensation, H.266/VVC splits the current CU into several  4 × 4 sub-CUs and calculates the MV of each sub-CU according to four-parameter or six-parameter affine models. The reference sub-CUs are then found through the corresponding MVs. By performing padding and fractional interpolation on the reference sub-CUs, the final predicted sub-CUs can be obtained. Finally, all the predicted sub-CUs are combined to obtain the predicted CU. In our method, we first obtain the minimum region containing all the reference sub-CUs and then pad the region and feed it to the neural network. Based on the corresponding position of the output image patch, the padded sub-CUs can be filtered by a DCTIF to obtain the predicted sub-CUs.

3.2. The Construction of an RBENN

Figure 6 depicts the overall framework of the RBENN. The network consists of 10 residual modules and 2 convolutional layers at the beginning and end. The network takes the padded reference CU before MC as input and passes it through a  3 × 3 convolution layer, which is followed by a rectified linear unit (ReLU) [26]. This process can be expressed as
F 1 = max ( 0 , W 0 · X ) ,
where  W 0 represents the convolution kernel of the first layer,  X is the input of the RBENN, and  F 1 is the output of the first layer.
The main body of the RBENN is composed of multiple residual convolution blocks (RCB), and the composition of the residual module [27] is also shown in Figure 6. It is composed of two convolutional layers, an ReLU, and a shortcut. The purpose of using a short cut is to make the network converged quickly. This process can be expressed as
F i , 1 = max ( 0 , W i , 1 · F i ) , F i , 2 = W i , 2 · F i , 1 , F i + 1 = F i + F i , 2 ,
where  1 i 10 i { 1 , 2 , , 10 } F i is the input of the i-th module,  W i , 1 is the first convolution kernel of the i-th module, and  W i , 2 is the second convolution kernel of the i-th module,  F i + 1 is the output of the i-th module and also the input of the  ( i + 1 ) -th module.
Finally, a  3 × 3 × 1 convolution kernel is used to compress all feature maps into a residual image, which is added to the input image to obtain the final output:
F 12 = W 12 · F 11 , Y = F 12 + X ,
where  Y is the output,  F 11 is the output of the last RCB, and  W 12 is the last convolution kernel.
The number of channels of convolution kernels in the whole network is 64, except in the last layer where it is 1.

3.3. Dataset for Training

In video coding, quantization will cause an irreversible loss of high-frequency image details. The aim of an RBENN is to restore the high frequency details of the image according to the low frequency information of the image.
In H.266/VVC, the relationship between different CUs is shown as
C U p r e = D C T I F ( C U r e f ) , C U r e s = C U o r i C U p r e , C U r e c = C U p r e + T 1 ( Q 1 ( Q ( T ( C U r e s ) ) ) ) ,
in which,  C U r e c C U p r e C U r e f C U o r i C U r e s represent the reconstruction of the current CU at the decoder, the predicted CU, the reference CU, the original CU, and the residual CU, respectively. The reference CU is the CU found on a reference frame according to MV. The predicted CU is obtained from the reference CU through the interpolation filter. The residual CU is transformed  T ( · ) , quantized  Q ( · ) , inversely quantized  Q 1 ( · ) , and inversely transformed  T 1 ( · ) to obtain the reconstructed CU.
In our proposed method, with the reference CU as input, the  R B E N N ( ) enhances the quality of the reference CU to replace the initial reference CU:
C U p r e = D C T I F ( R B E N N ( C U r e f ) ) .
For easy implementation, the input of the training data is the predicted CU, which is extracted from the decoder directly, and the label of the training data is the original CU, which is extracted from the original video sequence.
The specific details are shown in Figure 7, where the purple dots represent the integer pixel position of the picture. The current CU in the position  ( X , Y ) of the current frame finds the reference CU in the reference frame  ( X + I m v x , Y + I m v y ) position through MV, where  I m v represents the integer displacement of MV and  F m v is the fractional displacement. The reference CU is then filtered to obtain the predicted CU.
We used 650 video sequences to generate the training data. These 650 video sequences came from the dataset of BVI-DVC [28]. The resolutions were  3840 × 2176 1920 × 1088 960 × 544 , and  480 × 272 . To better adapt to video contents with different resolutions, all these video sequences were encoded using VTM-9.3 with LDP configuration [29]. The quantization parameter was 22. The first 32 frames of each sequence were encoded, the predicted CU of the integer MV was obtained from the compressed bitstream, and the original CU was obtained from the video sequence. The dataset contained  16 × 16 32 × 32 , and  64 × 64 CUs for training RBENN1, RBENN2, and RBENN3, respectively. There were about 10,000,000 blocks for training RBENN1, 2,000,000 blocks for training RBENN2, and 300,000 blocks for training RBENN3.

3.4. Training Strategy

Specifically, given a collection of n training samples, ( x i y i ),  i { 1 , 2 , , n } , the mean squared error (MSE) was used to minimize the loss function. The loss function is formulated as
L ( Θ ) = 1 n i = 1 n | | F ( x i | Θ ) y i | | 2 ,
where  Θ is the parameters of the network. The proposed RBENN was trained using PyTorch 1.8.0 on an NVIDIA GeForce GTX 2080Ti GPU. The loss function was minimized by using a first-order gradient based optimization method called Adam [30]. A batch-mode learning method was adopted with a batch size of 64. The momentum of Adam optimization was set to 0.9 and the momentum 2 was set to 0.999. The base learning rate was initially set to decay 0.1 times per 60 epochs from 0.0001. The training lasted 240 epochs.

4. Experimental Results

4.1. Experiment Settings

The proposed method was implemented based on H.266/VVC reference software VTM-9.3. The proposed RBENN was used to improve the quality of all luma CUs above  16 × 16 before MC. The LDP configuration was tested in the experiment under the H.266/VVC common test conditions. We encoded class D video sequences during the test, where the first 32 frames of each video sequence were encoded and decoded. There was no overlap between the training video sequences and the tested video sequences. The QPs used in our experiments varied among 22, 27, 32, and 37. The BD rate, which represents the bit rate increment at the same reconstruction quality, was used to measure the rate-distortion (RD) performance. The negative BD rate (BD rate reduction) means there is a positive coding gain, while a positive BD rate means there is a negative gain. To evaluate the reconstruction quality, PSNR and SSIM were used. Therefore, there are two kinds of BD rates: a PSNR-based BD rate and an SSIM-based BD rate.

4.2. Compared with VTM 9.3

The experimental results are summarized in Table 1, and RD curves of the several video sequences are shown in Figure 8. We can see that the proposed method achieved, on average, a 1.35% PSNR-based BD rate reduction and a 2.28% SSIM-based BD rate reduction. We can also see that the RBENN can save −2.15% of the PSNR-based BD rate for BQSquare and −3.96% of the SSIM-based BD rate for BlowingBubbles, which contains severe aliasing and more noise, indicating that the RBENN is more efficient for anti-aliasing and noise reduction. Figure 9 compares the subjective quality of the CUs with and without the proposed method. We can see that the RBENN did not affect the low-frequency information; however, it affected the high-frequency information, thus improving the image quality. Finally, we also tested the complexity of our proposed RBENN, as shown in Table 2. Since the CU is divided iteratively in the video coding process, frequent calls to the neural network lead to a large increase in complexity. In the future, we will reduce the complexity by designing lightweight networks.

4.3. Result Analysis

Since the RBENN enhances the reference CU before MC, the performance of the RBENN can be more clearly represented by comparing the PSNR of the reference CU and the original CU before and after the RBENN. Table 3 shows the average gains of the CU with different sizes in the four video sequences. The PSNR gains were calculated as the average of all CUs. Considering the different sizes and numbers of each CU, the PSNR gains were calculated as
P S N R g a i n = W , H { 16 , 32 , 64 } [ ( W × H ) × P S N R W , H ] W , H { 16 , 32 , 64 } [ ( W × H ) × N U M W , H ] ,
where  ( W × H ) is the area of a CU,  P S N R W , H represents the sum of PSNR gains for each CU after the RBENN, and  N U M W , H denotes the total number of CUs. According to Table 3, we can see that there was a 0.052 dB PSNR gain for all CUs larger than  16 × 16 , on average. For BQSquare, the PSNR gain reached 0.110 dB, which was consistent with the coding results. We can also see that the RBENN performed better for CUs with sizes of { 16 × 16 32 × 16 16 × 32 16 × 64 64 × 16 }. We believe that this was because the training set used for training RBENN1 was large, leading to a higher generalization ability. Figure 10 visually displays the distribution of the PSNR gains. We can see that, in the RaceHorses the RBENN achieved significant negative gains for a few CUs. This is the reason why limited coding improvement was obtained for it. In the future, we will add flags in the coding process to further improve the coding efficiency.

5. Conclusions

We proposed a neural-network-based quality enhancement for the reference CU to improve the accuracy of inter prediction. The network takes the luma reference CU before motion compensation as input. The output quality-enhanced blocks of the network are then interpolated by a DCTIF to generate a fractional motion compensation block. The proposed method was also implemented for affine motion compensation. Experimental results demonstrated that the proposed method could achieve a 1.35% reduction in the BD rate on average for the luma component, compared with H.266/VVC. In the future, we will design a more effective neural network to improve the coding efficiency and lightweight neural networks to reduce the complexity, while preserving the coding efficiency.

Author Contributions

Conceptualization, Y.C., H.Y. and S.J.; methodology, Y.C., H.Y., S.J. and C.F.; software, Y.C.; validation, Y.C.; formal analysis, Y.C.; investigation, Y.C.; resources, H.Y., S.J. and C.F.; data curation, Y.C., S.J. and C.F.; writing—original draft preparation, Y.C.; writing—review and editing, Y.C., H.Y., S.J. and C.F.; visualization, Y.C. and S.J.; supervision, S.J. and C.F.; project administration, H.Y. and S.J.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China under Grants 62222110 and 62172259, the Taishan Scholar Project of Shandong Province (tsqn202103001), the open project program of the State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, under Grant VRLAB2021A01, the Natural Science Foundation of Shandong Province of China under Grant (ZR2022ZD38), the Central Guidance Fund for Local Science and Technology Development of Shandong Province under Grant YDZX2021002, and the Major Scientific and Technological Innovation Project of Shandong Province under Grant 2020CXGC010109.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RBENNreference-block-enhancement neural network;
LDPlow-delay P;
MCmotion compensation;
MEmotion estimation;
DCTIFdiscrete cosine transform interpolation filter;
CUcoding unit;
BD rateBjøntegaard delta rate;
PSNRpeak signal to noise ratio.

References

  1. Rao, K.R.; Ahmed, N.; Natarajan, T. Discrete Cosine Transfom. IEEE Trans. Comput. 1974, 23, 90–93. [Google Scholar]
  2. Delogne, P.; Cuvelier, L.; Maison, B.; Van Caillie, B.; Vandendorpe, L. Improved interpolation, motion estimation, and compensation for interlaced pictures. IEEE Trans. Image Process. 1994, 5, 482–491. [Google Scholar] [CrossRef]
  3. Bjϕntegaard, G. Calculation of Average PSNR Differences between RD-Curves. ITU-T SG.16 Q.6 VCEG-M33. 2001. Available online: https://www.itu.int/wftp3/av-arch/video-site/0104_Aus/VCEG-M33.doc (accessed on 18 December 2022).
  4. Ugur, K.; Alshin, A.; Alshina, E.; Bossen, F.; Han, W.; Park, J.; Lainema, J. Interpolation filter design in HEVC and its coding efficiency—Complexity analysis. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 1704–1708. [Google Scholar]
  5. Yan, N.; Liu, D.; Li, H.; Wu, F. A Convolutional Neural Network Approach for Half-pel Interpolation in Video Coding. In Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), Baltimore, MD, USA, 28–31 May 2017; pp. 1–4. [Google Scholar]
  6. Yan, N.; Liu, D.; Li, H.; Li, B.; Li, L.; Wu, F. Convolutional Neural Network-Based Fractional-Pixel Motion Compensation. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 840–853. [Google Scholar] [CrossRef]
  7. Yan, N.; Liu, D.; Li, H.; Li, B.; Li, L.; Wu, F. Invertibility-driven Interpolation Filter for Video Coding. IEEE Trans. Image Process. 2019, 28, 4912–4925. [Google Scholar] [CrossRef]
  8. Xia, S.; Yang, W.; Hu, Y.; Ma, S.; Liu, J. A Group Variational Transformation Neural Network for Fractional Interpolation of Video Coding. In Proceedings of the 2018 Data Compression Conference, Snowbird, UT, USA, 27–30 March 2018; pp. 127–136. [Google Scholar]
  9. Liu, J.; Xia, S.; Yang, W.; Li, M.; Liu, D. One-for-All: Grouped Variation Network-Based Fractional Interpolation in Video Coding. IEEE Trans. Image Process. 2019, 28, 2140–2151. [Google Scholar] [CrossRef]
  10. Huo, S.; Liu, D.; Wu, F.; Li, H. Convolutional Neural Network-Based Motion Compensation Refinement for Video Coding. In Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; pp. 1–4. [Google Scholar]
  11. Wang, Y.; Fan, X.; Xiong, R.; Zhao, D.; Gao, W. Neural Network-Based Enhancement to Inter Prediction for Video Coding. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 826–838. [Google Scholar] [CrossRef]
  12. Pham, C.D.; Zhou, J. Deep Learning-Based Luma and Chroma Fractional Interpolation in Video Coding. IEEE Access 2019, 7, 112535–112543. [Google Scholar] [CrossRef]
  13. Prette, N.; Valsesia, D.; Bianchi, T. Deep Multiframe Enhancement for Motion Prediction in Video Compression. In Proceedings of the IEEE International Conference on Electronics, Circuits, and Systems (ICECS), Dubai, United Arab Emirates, 28 November–1 December 2021; pp. 1–6. [Google Scholar]
  14. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wieg, T. Overview of the High Efficiency Video Coding (HEVC) Standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  15. Murn, L.; Blasi, S.; Smeaton, A.F.; O’Connor, N.E.; Mrak, M. Interpreting CNN For Low Complexity Learned Sub-Pixel Motion Compensation In Video Coding. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 798–802. [Google Scholar]
  16. Murn, L.; Blasi, S.; Smeaton, A.F.; Mrak, M. Improved CNN-Based Learning of Interpolation Filters for Low-Complexity Inter Prediction in Video Coding. IEEE Open J. Signal Process. 2021, 2, 453–465. [Google Scholar] [CrossRef]
  17. Galpin, F.; Bordes, P.; Dumas, T.; Nikitin, P.; Le Leannec, F. Neural Network based Inter bi-prediction Blending. In Proceedings of the 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany, 5–8 December 2021; pp. 1–5. [Google Scholar]
  18. Jin, D.; Lei, J.; Peng, B.; Li, W.; Ling, N.; Huang, Q. Deep Affine Motion Compensation Network for Inter Prediction in VVC. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3923–3933. [Google Scholar] [CrossRef]
  19. Katayama, T.; Song, T.; Shimamoto, T.; Jiang, X. Reference Frame Generation Algorithm using Dynamical Learning PredNet for VVC. In Proceedings of the 2021 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 10–12 January 2021; pp. 1–5. [Google Scholar]
  20. Bross, B.; Chen, J.; Ohm, J.R.; Sullivan, G.J.; Wang, Y.K. Developments in International Video Coding Standardization After AVC, With an Overview of Versatile Video Coding (VVC). Proc. IEEE 2021, 109, 1463–1493. [Google Scholar] [CrossRef]
  21. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  22. Dai, Y.; Liu, D.; Wu, F. A Convolutional Neural Network Approach for Post-Processing in HEVC Intra Coding. In Proceedings of the 2017 International Conference on Multimedia Modeling (MMM), Reykjavik, Iceland, 4–6 January 2017; pp. 28–39. [Google Scholar]
  23. Sun, D.; Yang, X.; Liu, M.; Kautz, J. PWC-Net: CNNs for Optical Flow Using Pyramid, Warping, and Cost Volume. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8934–8943. [Google Scholar]
  24. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  25. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  26. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Ma, D.; Zhang, F.; Bull, D.R. BVI-DVC: A Training Database for Deep Video Compression. IEEE Trans. Multimed. 2022, 24, 3847–3858. [Google Scholar]
  29. Liu, S.; Segall, A.; Alshina, E.; Liao, R. JVET Common Test Conditions and Evaluation Procedures for Neural Network-Based Video Coding Technology. Joint Video Experts Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29, Document no. JVET-T2006. January 2021. Available online: https://www.itu.int/wftp3/av-arch/jvet-site/2021_01_U_Virtual/JVET_Notes_d1.docx (accessed on 18 December 2022).
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 2015 International Conference on Learning Representations(ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
Figure 1. Inter prediction.
Figure 1. Inter prediction.
Applsci 13 02795 g001
Figure 2. Relationship between the current CU, reference CU, and predicted CU; the block with green dots represents the integer pixel position.
Figure 2. Relationship between the current CU, reference CU, and predicted CU; the block with green dots represents the integer pixel position.
Applsci 13 02795 g002
Figure 3. VVC coding framework.
Figure 3. VVC coding framework.
Applsci 13 02795 g003
Figure 4. A diagram of RBENN integrated into H.266/VVC.
Figure 4. A diagram of RBENN integrated into H.266/VVC.
Applsci 13 02795 g004
Figure 5. The detail of an RBENN integrated before MC.
Figure 5. The detail of an RBENN integrated before MC.
Applsci 13 02795 g005
Figure 6. Framework of an RBENN.
Figure 6. Framework of an RBENN.
Applsci 13 02795 g006
Figure 7. Illustration of the process of generating training data for training the RBENN.
Figure 7. Illustration of the process of generating training data for training the RBENN.
Applsci 13 02795 g007
Figure 8. RD curves of class D video sequences under LDP.
Figure 8. RD curves of class D video sequences under LDP.
Applsci 13 02795 g008
Figure 9. Quality enhancement results on reference CU performed by the RBENN.
Figure 9. Quality enhancement results on reference CU performed by the RBENN.
Applsci 13 02795 g009
Figure 10. Number of CUs with different gains.
Figure 10. Number of CUs with different gains.
Applsci 13 02795 g010
Table 1. Results of Our RBENN Compared to VTM9.3 under LDP.
Table 1. Results of Our RBENN Compared to VTM9.3 under LDP.
Class DQPVTM-9.3Our MethodBD Rate Y
kbpsY PSNRY SSIMkbpsY PSNRY SSIMPSNRSSIM
BasketballPass22462.0241.770.9702460.6141.790.9704−1.49%−2.25%
27212.1237.830.9402210.3937.880.9407
32105.2234.520.8965104.8734.590.8984
3754.5031.660.842054.3431.730.8434
BQSquare22938.1538.650.9435927.0638.700.9439−2.15%−2.01%
27346.2934.580.8972342.8434.650.8986
32165.9431.570.8381166.2931.660.8401
3789.2328.860.766089.6528.940.7676
BlowingBubbles221086.8938.350.96221085.6738.370.9637−1.14%−3.96%
27428.8834.620.9263427.4734.670.9280
32192.4931.640.8758192.4631.680.8778
3790.4429.030.834189.9429.030.8371
RaceHorses221307.5639.170.96141306.5039.210.9619−0.61%−0.91%
27582.0834.730.9144579.6434.740.9146
32273.9531.230.8471272.8031.240.8481
37134.7228.490.7725134.4828.510.7719
Average −1.35%−2.28%
Table 2. Complexity of our method.
Table 2. Complexity of our method.
Class DLDP
EncTimeDecTime
BasketballPass70,977%17,959%
BQSquare64,455%30,334%
BlowingBubbles58,523%26,708%
RaceHorses47,806%26,507%
Average56,653%24,021%
Table 3. PSNR gain for RBENN for different CUs.
Table 3. PSNR gain for RBENN for different CUs.
Class DAverage (CNN−DCT) PSNR for Each Block
16 × 16 16 × 32 32 × 16 32 × 32 16 × 64 64 × 16 32 × 64 64 × 32 64 × 64 PSNR Gain
BasketballPass0.1250.0880.0660.0060.0800.0090.0120.0280.0050.036
BQSquare0.1750.1770.1790.0510.2030.1500.0680.0380.0550.110
BlowingBubbles0.0670.0500.0400.0110.0290.0280.0140.0080.0000.030
RaceHorses0.0720.0470.0140.0030.027−0.0400.007−0.042−0.150.030
Average0.1100.0910.0750.0180.0850.0370.0250.008−0.0230.052
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chu, Y.; Yuan, H.; Jiang, S.; Fu, C. Neural Network-Based Reference Block Quality Enhancement for Motion Compensation Prediction. Appl. Sci. 2023, 13, 2795. https://0-doi-org.brum.beds.ac.uk/10.3390/app13052795

AMA Style

Chu Y, Yuan H, Jiang S, Fu C. Neural Network-Based Reference Block Quality Enhancement for Motion Compensation Prediction. Applied Sciences. 2023; 13(5):2795. https://0-doi-org.brum.beds.ac.uk/10.3390/app13052795

Chicago/Turabian Style

Chu, Yanhan, Hui Yuan, Shiqi Jiang, and Congrui Fu. 2023. "Neural Network-Based Reference Block Quality Enhancement for Motion Compensation Prediction" Applied Sciences 13, no. 5: 2795. https://0-doi-org.brum.beds.ac.uk/10.3390/app13052795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop