Joint Reference Frame Synthesis and Post Filter Enhancement for Versatile Video Coding
Abstract
This paper presents the joint reference frame synthesis (RFS) and post-processing filter enhancement (PFE) for Versatile Video Coding (VVC), aiming to explore the combination of different neural network-based video coding (NNVC) tools to better utilize the hierarchical bi-directional coding structure of VVC. Both RFS and PFE utilize the Space-Time Enhancement Network (STENet), which receives two input frames with artifacts and produces two enhanced frames with suppressed artifacts, along with an intermediate synthesized frame. STENet comprises two pipelines, the synthesis pipeline and the enhancement pipeline, tailored for different purposes. During RFS, two reconstructed frames are sent into STENet’s synthesis pipeline to synthesize a virtual reference frame, similar to the current to-be-coded frame. The synthesized frame serves as an additional reference frame inserted into the reference picture list (RPL). During PFE, two reconstructed frames are fed into STENet’s enhancement pipeline to alleviate their artifacts and distortions, resulting in enhanced frames with reduced artifacts and distortions. To reduce inference complexity, we propose joint inference of RFS and PFE (JISE), achieved through a single execution of STENet. Integrated into the VVC reference software VTM-15.0, RFS, PFE, and JISE are coordinated within a novel Space-Time Enhancement Window (STEW) under Random Access (RA) configuration. The proposed method could achieve -7.34%/-17.21%/-16.65% PSNR-based BD-rate on average for three components under RA configuration.
Index Terms:
neural network-based video coding, Versatile Video Coding (VVC), inter prediction, post-processing filter.I Introduction
Over the past few decades, numerous video coding standards have emerged to compress the ever-expanding volume of video data. The latest video coding standard Versatile Video Coding (VVC) was released in July 2020 by the Joint Video Experts Team (JVET), which was founded by two international standardization organizations, ITU-T Video Coding Experts Group (VCEG) and ISO/IEC Moving Picture Experts Group (MPEG)[1]. Compared to the previous video standard High Efficiency Video Coding (HEVC), VVC reaches approximately 50% bit rate reduction for nearly equal video quality[2]. Moreover, VVC is more versatile for certain types of video content, such as ultra-high resolution videos, 360° videos, and gaming content videos[1].
With the development of deep learning, neural network-based video coding (NNVC) has made tremendous progress. The existing works can be classified into two categories: deep schemes and deep tools[3]. The deep schemes are built primarily upon end-to-end deep networks as encoders and decoders. Deep schemes have garnered significant attention due to their promising prospects[4, 5, 6]. On the other hand, the deep tools (NNVC tools) shall be used within traditional coding schemes or together with traditional coding tools. These NNVC tools can provide additional performance enhancements when integrated into traditional coding schemes.
Inter prediction is a crucial part of block-based hybrid coding schemes that reduce temporal redundancy in video sequences. During inter prediction, the to-be-coded frame will select frames from reference picture lists (RPLs) as reference frames for motion estimation (ME) and motion compensation (MC)[7]. Reference frames much more similar to the to-be-coded frame will lead to smaller residuals in the bitstream. This characteristic inspires some works to enhance inter prediction by synthesizing reference frames[8, 9, 10, 11]. During inter prediction, different kinds of reference frame synthesis networks are required for bi-directional prediction and uni-directional prediction. Specifically, interpolation networks enhance bi-directional prediction, while extrapolation networks improve uni-directional prediction.
Meanwhile, block partitioning and quantization in block-based hybrid coding schemes lead to artifacts in reconstructed frames, including block artifacts, ringing artifacts, and over-smoothing. In VVC, various in-loop filters such as deblocking filter (DBF), sample adaptive offset (SAO), adaptive loop filtering (ALF), cross-component adaptive loop filtering (CC-ALF), and luma mapping with chroma scaling (LMCS) are employed to mitigate these artifacts[12]. Nowadays, some work has demonstrated that combining neural network-based in-loop filters with handcrafted in-loop filters can lead to significant coding improvements[13, 14]. Meanwhile, some researchers have introduced neural network-based post-processing filters designed to improve video quality[15, 16, 17, 18]. Different from neural network-based in-loop filters, neural network-based post-processing filters can apply to any video standard without any codec modifications.
In this paper, we propose a novel approach: joint reference frame synthesis (RFS) and post-processing filter enhancement (PFE) for VVC, aiming to explore the combination of different NNVC tools to better utilize the hierarchical bi-directional coding structure of VVC. By combining RFS and PFE, the proposed method achieves space-time coupled enhancement among neighboring frames. In VVC, Random Access (RA) configuration defines a hierarchical bi-directional coding structure, which effectively utilizes temporal information in contrast to both Low Delay P (LDP) and Low Delay B (LDB) configurations. Considering this, the paper focuses on RA configuration to better utilize bi-directional temporal information for space-time coupled enhancement. The overall framework is shown in Fig.1. Both RFS and PFE utilize the Space-Time Enhancement Network (STENet), illustrated in Fig.2. STENet takes two compressed frames with artifacts and distortions, fully leveraging the space-time information to generate an intermediate synthesized frame and two enhanced frames with reduced artifacts. STENet comprises two pipelines: the synthesis pipeline for RFS and the enhancement pipeline for PFE. During RFS, two reconstructed frames from the decoded picture buffer (DPB) are input into STENet’s synthesis pipeline to synthesize an intermediate frame, treated as the virtual reference frame, and inserted into two RPLs. Meanwhile, during PFE, two reconstructed frames are selected from DPB and input into STENet’s enhancement pipeline to alleviate their artifacts and distortions, resulting in higher-quality final output frames. To reduce inference complexity, we propose joint inference of RFS and PFE (JISE), achieved through a single execution of STENet for RFS and PFE. Additionally, to effectively manage RFS, PFE, and JISE under RA configuration, we introduce the Space-Time Enhancement Window (STEW), illustrated in Fig.3. For training, we adopt a joint training strategy for STENet, simultaneously optimizing both the synthesis and enhancement pipelines. This simplifies training and thoroughly considers the performance impact between RFS and PFE. The proposed method is integrated into the VVC reference software VTM-15.0, with experimental results showcasing significant performance enhancements in VVC under RA configuration.
The remainder of the paper is organized as follows: Section II reviews the background and related work. Details of our proposed method are given in Section III. Section IV presents our experiments and analysis. Finally, the paper is concluded in Section V.
II Related Work
II-A Neural network-based Video Coding Tools
NNVC tools shall be utilized within traditional coding schemes or together with traditional coding tools. There have been several techniques thoroughly researched, including in-loop filter[13, 14], post-processing filter[15, 16, 17, 18], intra prediction[19, 20], inter prediction[21, 8, 9, 10, 11] and super resolution[22, 23]. More research about NNVC tools can be found in [3, 24]. Meanwhile, JVET has placed significant emphasis on the exploration and standardization of NNVC tools. JVET has established Ad-Hoc Group 11 (AHG11)[25], which intends to study the performance, complexity, and training processes of NNVC tools. Some works by JVET experts have achieved significant improvements[26]. In addition to well-designed networks and usage strategies, training material needs to include diverse content covering different formats and texture types. JVET NNVC exploration activities have utilized BVI-DVC[27] and TVD[28] as training materials. As for neural network inference, F. Galpin et al.[29] propose the Small Ad-hoc Deep-Learning Library (SADL), a header-only small C++ library for the inference of NNVC tools. The objective of SADL is to offer a straightforward framework that is dependency-free and compliant with the software development policy of JVET.
II-B Neural network-based Post-processing Filter Enhancement
PFE seeks to reduce artifacts introduced by block partitioning and quantization, including block artifacts, ringing artifacts, and over-smoothing. Compared to in-loop filters, PFE does not require any modification to the codec, making it applicable to any video standard. Recently, neural networks have made significant contributions to image and video restoration[30, 31, 32], which has catalyzed the vibrant development of neural network-based PFE. Zhang et al.[15] introduce a neural network-based post-processing network for VVC in RA configuration. To enhance the robustness, they trained five different models for five QP groups to accommodate input frames of varying qualities. In [16], the MFRNet, comprising four multi-level feature review residual dense blocks (MFRBs), is proposed for post-processing and in-loop filtering. Their experimental results demonstrated that MFRNet is more advantageous for post-processing and in-loop filtering than other popular network structures. Santamaria et al.[17] propose a content-adaptive convolution neural network post-processing filter, which can be trained offline on general video sequences and later fine-tuned on the test video sequence.
However, the majority of PFE methods solely rely on the spatial information of the current frame, disregarding the valuable temporal information from neighboring frames. For instance, while the current frame may contain numerous artifacts on the same object, there might be only a few artifacts present in another frame. In such cases, leveraging the temporal information from another frame can help mitigate the artifacts.
II-C Neural network-based Reference Frame Synthesis
RFS aims to synthesize frames that highly similar to the to-be-coded frame to enhance inter prediction. These synthesized frames are inserted into RPLs and treated as the virtual reference frames. Some RFS methods enhance bi-direction prediction by utilizing Video Frame Interpolation (VFI) networks to generate intermediate virtual reference frames. For instance, Zhao et al.[8] interpolate reconstructed frames as the prediction frame, requiring the encoder to choose between traditional and emerging results using rate-distortion optimization (RDO) at the coding unit (CU) level. Guo et al.[9] synthesize a new reference frame from two-sided previously reconstructed frames and then append the synthesized frame to the RPL, which improves the compression performance of HM-16.20. Furthermore, some studies have explored using video frame extrapolation networks to enhance uni-direction prediction. Huo et al.[33] utilize a neural network-based extrapolation network to synthesize reference frames, supporting LDP and LDB configurations. Their approach aligns reference frames through block-based motion estimation and compensation, followed by extrapolation from the aligned frames. Jia et al.[11] present a deep reference frame generation method in which interpolation and extrapolation networks are utilized to enhance bi-directional prediction and uni-directional prediction, respectively.
Although RFS methods have been developed for many years, there is still room for improvement. Many existing methods simply utilize interpolation or extrapolation networks to directly synthesize reference frames. These approaches often overlook the distortions and artifacts present in the input reconstructed frames, which can negatively impact the quality of the synthesized frames. Furthermore, numerous RFS methods fail to consider the interactions with other NNVC tools. When other NNVC tools are active, the performance of the RFS tool may be significantly hindered[34].
III Proposed Method
In this section, we begin by outlining the motivation behind our proposed method. Next, we introduce STENet, as depicted in Fig.2. Subsequently, we delve into the investigation of RFS, PFE, and JISE within the context of VVC, as illustrated in Fig.1. Finally, we present the inference details of our proposed method.
III-A Motivation
Despite years of development, there is still room for improvement in NNVC. Initially, most NNVC tools overlook the importance of incorporating space-time information across neighboring frames, eg., PFE methods primarily concentrate on spatial information. Moreover, each NNVC tool is trained independently. Different NNVC tools often require distinct training materials and strategies, leading to substantial human and computational resources. Besides resource consumption, the isolated training of different NNVC tools may result in significant performance overlap. Additionally, there has been limited research exploring the feature reuse among different NNVC tools. By reusing features, different NNVC tools can reduce the complexity of feature extraction and achieve accelerated inference.
To address the aforementioned issues, we propose a joint utilization of RFS and PFE, which contain one joint network, one joint training strategy, and one joint inference. The proposed method realizes the space-time coupled enhancement among neighboring frames. Firstly, we introduce the STENet, which serves the dual purpose of RFS and PFE. The STENet contains a synthesis pipeline that generates a virtual reference frame for RFS, and an enhancement pipeline that improves the quality of two input frames for PFE. Secondly, we adopt a joint training strategy for STENet, with both the synthesis and enhancement pipelines optimized simultaneously. This approach simplifies the training process and demonstrates superior performance compared to the isolated training of two pipelines. Lastly, we propose a joint inference mechanism, JISE, which enables simultaneous inference of RFS and PFE. Unlike independent inference, the JISE significantly reduces the overall inference time by inferring RFS and PFE concurrently, thereby improving efficiency.
III-B Space-Time Enhancement Network
The STENet receives two input frames and , and produces two enhanced frames and with suppressed artifacts, along with an intermediate synthesized frame . This process can be described as:
(1) |
where represents the STENet. As depicted in Fig.2, the STENet employs a bidirectional recurrent architecture to realize global information propagation. In this bidirectional recurrent setup, both frames and features are propagated in two opposite directions. This allows the input frames to gather information from any other frames, and the intermediate synthesized frame to leverage information from the two input frames. The two recurrent processes are divided into the forward and backward branches depending on the direction of information propagation.
III-B1 Optical Flow Estimation Module
The optical flow estimation module is illustrated in Fig.2(b). To achieve feature alignment between two input frames, it’s essential to estimate the optical flows and between them. Additionally, for interpolating an intermediate frame from two input frames, estimation of the optical flows and from the intermediate moment is required. Inspired by [35], which proposes an effective and precise optical flow reusing strategy, we adopt the same strategy to reduce the complexity involved in optical flow estimation.
To begin with, we utilize the IFRNet[36] as the optical flow estimator due to its fast inference speed. Since IFRNet requires inputs in RGB format, we first upsample the U and V channels of each input frame in YUV420 format and then perform a color space transformation to convert them into RGB format. The reformatted input frames are then fed into IFRNet, which estimates the optical flows and . This process can be summarized as follows:
(2) |
where represents the intermediate optical flow estimator. Next, following the approach described in [35], we adopt a similar strategy to generate the optical flows and based on the previously estimated optical flows and . This can be expressed as:
(3) |
where represents the optical flow reusing strategy from [35]. At this stage, all the optical flows have been estimated. These resulting optical flows will be utilized for alignment in the bidirectional recurrence, which will be explained in detail in the subsequent section.
III-B2 Network Architecture
Inspired by [37], we downsample the input frames and the optical flows to reduce the computational cost. For each input frame in YUV420 format, the Y component undergoes a PixelUnshuffle operation[38] with a scale factor of 2, followed by a concatenation operation with the U and V components to generate a six-channel tensor. The optical flows, detailed in Section III-B1, are downsampled accordingly. These designs enable most of the computations to be performed in the low-resolution feature space, thereby greatly enhancing efficiency. It’s worth noting that, instead of downsampling the input frames to estimate optical flows as mentioned in [37], performing high-resolution optical flow estimation and then downsampling optical flows will yield more accurate low-resolution optical flows.
In the backward branch, information from the succeeding state is transmitted to both the preceding and intermediate states. A weight-shared residual refinement module is utilized to refine features at each state. The feature extraction process of the succeeding state is described as:
(4) |
where represents the weight-shared feature extraction module and denotes the channel concatenation operation. To align the input channel of in each state, the tensor filled with zeros is treated as a placeholder. Then, the preceding state receives knowledge from the succeeding feature:
(5) |
where denotes the flow warping operation. Next, the knowledge from the succeeding state is propagated to the intermediate state as prior knowledge for the frame synthesis, represented as:
(6) |
Up to now, via backward propagation, we acquire three features in distinct states, which are intended for further refinement within the forward branch.
The forward branch shares the same idea with the backward branch. Both the succeeding and intermediate states will receive knowledge from the preceding state. The specific processes at three different states are as follows:
(7) |
(8) |
(9) |
where represents the weight-shared residual refinement module in the forward branch. So far, we have extracted features in three states. A weight-shared convolution layer was employed for each state for reconstruction, yielding a six-channel tensor. To get the output frame in YUV420 format, the first four channels of the reconstructed feature undergo a PixelShuffle operation[38] with a scale factor of to obtain the Y component, and the last two channels are U and V components, respectively.
As depicted in Fig.2(a), STENet comprises two pipelines: the enhancement pipeline and the synthesis pipeline. These pipelines can operate jointly or independently based on routing selection. The enhancement pipeline takes two input frames and as input to enhance each other, resulting in two enhanced frames and :
(10) |
where the represents the enhancement pipeline. Meanwhile, the synthesis pipeline takes two input frames and for frame synthesis, producing an intermediate synthesized frame :
(11) |
where the denotes the synthesis pipeline. Additionally, the enhancement pipeline can be utilized for PFE, the synthesis pipeline for RFS, and the entire STENet for JISE. Further details are provided in the subsequent section.
III-C Integration into VVC
The framework of our proposed method, as illustrated in Fig.1, integrates both RFS and PFE into the VVC reference software VTM-15.0. RFS aims to synthesize frames that highly similar to the to-be-coded frame to enhance inter prediction. These synthesized frames are inserted into RPLs and treated as the virtual reference frames. On the other hand, PFE operates outside the encoding and decoding loops to enhance the reconstructed frames. Meanwhile, to reduce computation costs, we propose JISE, in which the RFS and PFE are executed jointly.
III-C1 Space-Time Enhancement Window
To effectively coordinate RFS, PFE, and JISE under RA configuration, we introduce the STEW. As depicted in Fig.3(a), each STEW contains 8 consecutive frames. Within the current STEW, we determine whether to execute RFS, PFE, or JISE based on the POC of the current to-be-coded frame. The specific mode decision process is as follows:
(12) |
where denotes RFS, represents PFE, and indicates JISE.
In each STEW, the processes of RFS, PFE, and JISE are executed alternately, their execution sequence is constrained by the encoding order of the RA configuration. As illustrated in Fig.3(a), we annotate their execution sequence using symbols. For a specific symbol, the superscript denotes its order, while the subscript indicates whether it is RFS, PFE, or JISE. synthesizes an additional reference frame for the frame, followed by , which synthesizes an additional reference frame for the frame. Subsequently, synthesizes an additional reference frame for the frame and enhances the and frames. Finally, synthesizes an additional reference frame for the frame, and enhances the and frames. Further operations are not described in detail. In each STEW, the latter 7 frames will reference the synthesized reference frames generated by RFS or JISE, and all frames will be enhanced by PFE or JISE. Once the last frame is decoded and enhanced, the STEW will slide forward by 8 POC distances.
III-C2 Reference Frame Synthesis
DPB stores the reconstructed frames during encoding. Specific reconstructed frames are selected from DPB to be included in the RPL as reference frames. For bi-directional inter prediction, two separate RPLs are established in different directions for the to-be-coded frame. The encoder searches the best-matching blocks in reference frames as predictions of the current to-be-coded CU. Reference frames with more similar content to the to-be-coded frame can lead to smaller residual coding consumption in the bitstream. To enhance the bi-directional inter prediction in VVC, the JVET CTC [39] specifies RA configuration, which employs a hierarchical coding structure with 32 frames in a group of pictures (GOP).
Within each STEW, RFS synthesizes a virtual reference frame to enhance bi-directional inter prediction under RA configuration. As illustrated in Fig.3(a), the frames within a GOP are divided into 6 temporal layers in RA configuration, allowing the utilization of previously encoded frames in lower layers as references for the to-be-coded frames in higher layers. As shown in Fig.3(b), RFS selects two-sided reconstructed frames and as inputs to synthesize a virtual reference frame through the STENet’s synthesis pipeline, represented as:
(13) |
where represents STENet’s synthesis pipeline. Considering the degradation of frame synthesis performance over long temporal distances, the POC distance should be less than 4. In current STEW, the relationship between the POC distance and the to-be-coded frame’s POC is arranged as:
(14) |
As for the other frames, in which , we will provide detailed explanations in Section III-C4.
The synthesized frame is inserted into two RPLs (RPL0 and RPL1) of the current to-be-coded frame and treated as additional reference frames in two RPLs. The indexes of the synthesized frame in RPL0 and RPL1 are set to 1. Notably, the synthesized reference frame is utilized like other reference frames, thus ensuring that the encoder upholds consistent ME and MC procedures during inter prediction. RFS operates in both the encoding and decoding processes, eliminating the necessity to convey additional information into the bitstream.
III-C3 Post-processing Filter Enhancement
In block-based hybrid coding schemes, reconstructed frames often suffer from block artifacts, ringing artifacts, and over-smoothing. Those artifacts will degrade the perceptual quality of the reconstructed video. In most block-based hybrid coding schemes, several in-loop filters operate inside the decoder to alleviate artifacts in filtered frames. On the other hand, post-processing filters operate out of the encoding and decoding loop. This means that post-processing filters can be integrated into any video standard without any codec modifications. However, most post-processing filters only use information from the current to-be-coded frame, ignoring potentially useful information from neighboring frames.
Within each STEW, PFE filters the reconstructed frames to mitigate their artifacts and distortions. PFE simultaneously improves two input frames, enabling each frame to leverage its own spatial information and the temporal information from another frame to improve its quality. As shown in Figure 3(c), PFE takes two reconstructed frames and as inputs and uses the STENet’s enhancement pipeline to improve them. If the POC distance between the two reconstructed frames is either too short or too long, the reconstructed frame will not be able to effectively get complementary information from other frames. Therefore, we set the POC distance between the two reconstructed frames to be 2. The process can be described as follows:
(15) |
where represents STENet’s enhancement pipeline, and are two enhanced frames, and is the POC of the current to-be-coded frame. In other words, the and reconstructed frames are filtered simultaneously, while the and reconstructed frames are filtered simultaneously. As for the other frames in the current STEW, we will provide detailed explanations in Section III-C4. The enhanced frames are considered as the final output frames and are not used as reference frames.
If the reconstructed frames exhibit minimal artifacts, PFE may degrade the quality of the enhanced frames. Hence, we offer the flexibility to enable or disable PFE at the block level. We utilize PSNR metrics to assess the similarity between the original, filtered, and reconstructed blocks. If the original block closely resembles the filtered block more than the reconstructed block, we activate PFE for that block. In contrast, we will turn off the PFE in this block. Switch flags indicating the PFE status are subsequently embedded into the bitstream and transmitted to the decoder.
III-C4 Joint Inference of RFS and PFE
In our design, the PFE and RFS are continuous in some positions. For instance, once the and frames in the current STEW are decoded, PFE can utilize STENet’s enhancement pipeline to filter them. After PFE, RFS can utilize STENet’s synthesis pipeline to synthesize a virtual reference frame for the frame. This inspires us to introduce the JISE, in which RFS and PFE are executed jointly. Instead of utilizing RFS and PFE independently, JISE can reduce network inference complexity by reusing features between two pipelines of STENet.
For the first three frames in the current STEW, we can input the and reconstructed frames into STENet to filter them and synthesize an intermediate frame. The synthesized frame is treated as the virtual reference frame for the frame. As for the , , and frames in the current STEW, the process is the same. The process can be described as follows:
(16) |
where represents the STENet. The enhanced frames and are treated as final output frames after determining whether to keep enhanced blocks or not at the block level. The synthesized frame is treated as a virtual reference frame, which will be inserted into two RPLs of the current to-be-coded frame .
The synthesis and enhancement pipelines reuse lots of modules. When RFS and PFE are executed independently, those reused modules will be inferred twice. In contrast, those reused modules will infer only once within JISE. That is why the proposed JISE can reduce network inference complexity.
III-D Inference Details
III-D1 Block-based Network Inference
During network inference, large input sizes can lead to excessive memory consumption, especially for (4K) resolution inputs. To solve this problem, we divide the input frames into blocks and work on them separately. However, using small blocks can limit the amount of spatial and temporal information captured, leading to a notable drop in STENet performance. Therefore, we use different block sizes for different resolutions. For JVET NNVC common test condition (CTC) class D[40] input frames, we use blocks with an extra padding of . For input frames from other CTC classes, we use blocks with an extra padding of . It is worth mentioning that the activation or deactivation decisions for the PFE are also made at the block level as previously described.
III-D2 Temporary Buffer for Enhanced Frames
After PFE, the enhanced frames are temporarily stored in a buffer and released as output frames once all frames within the current STEW are enhanced. The temporary buffer ensures that the inputs of RFS must be the reconstructed frames without being enhanced by PFE.
Without this design, frames enhanced by PFE will be utilized for subsequent RFS. For example, as depicted in Fig.3(a), following the process, the frame is enhanced. This enhanced frame is subsequently utilized as input during the process, potentially degrading RFS performance. Specifically, during training, STENet takes reconstructed frames with compression artifacts as inputs. However, during inference, if the inputs of STENet are enhanced frames with suppressed artifacts, the quality of the synthesized frame may decline. Moreover, the synthesized frame may not serve as an effective reference for the to-be-coded frame.
IV Experiments
This section provides a detailed elucidation of the experimental setup and results. It includes details about the training data generation and the joint training strategy. Furthermore, we present both subjective and objective analyses of the proposed method, along with ablation experiments. Additionally, we conduct comparisons between the proposed method and existing methodologies.
Class | Sequence | BD-rate | |||||
---|---|---|---|---|---|---|---|
Y-PSNR | U-PSNR | V-PSNR | Y-MSIM | U-MSIM | V-MSIM | ||
Class A1 | Tango2 | -10.80% | -25.75% | -24.47% | -11.91% | -29.64% | -27.15% |
FoodMarket4 | -4.88% | -10.89% | -11.76% | -5.43% | -12.86% | -13.42% | |
Campfire | -2.17% | -0.02% | -4.72% | -3.73% | -1.06% | -8.47% | |
Average | -5.95% | -12.22% | -13.65% | -7.02% | -14.52% | -16.35% | |
Class A2 | CatRobot | -9.17% | -23.55% | -19.90% | -10.04% | -20.54% | -16.87% |
DaylightRoad2 | -10.42% | -19.19% | -13.69% | -10.30% | -20.59% | -14.43% | |
ParkRunning3 | -2.83% | -4.49% | -3.19% | -4.65% | -6.74% | -5.36% | |
Average | -7.48% | -15.74% | -12.26% | -8.33% | -15.96% | -12.22% | |
Class B | MarketPlace | -5.75% | -19.58% | -16.86% | -6.33% | -20.71% | -15.65% |
RitualDance | -6.60% | -13.23% | -17.44% | -6.73% | -14.22% | -17.97% | |
Cactus | -6.25% | -15.28% | -8.08% | -7.15% | -18.09% | -12.81% | |
BasketballDrive | -7.67% | -17.90% | -15.42% | -8.73% | -19.65% | -21.36% | |
BQTerrace | -5.37% | -15.87% | -11.91% | -5.47% | -16.54% | -14.46% | |
Average | -6.33% | -16.37% | -13.94% | -6.88% | -17.84% | -16.63% | |
Class C | BasketballDrill | -7.92% | -16.88% | -17.18% | -9.97% | -18.48% | -21.31% |
BQMall | -10.61% | -26.10% | -25.79% | -12.62% | -25.81% | -24.69% | |
PartyScene | -6.39% | -14.89% | -13.43% | -6.25% | -12.84% | -12.16% | |
RaceHorses | -7.59% | -23.38% | -26.08% | -9.54% | -21.43% | -23.31% | |
Average | -8.13% | -20.31% | -20.62% | -9.59% | -19.64% | -20.37% | |
Class D | BasketballPass | -10.00% | -26.44% | -23.94% | -10.53% | -28.54% | -24.76% |
BQSquare | -10.54% | -13.39% | -21.89% | -7.35% | -15.06% | -19.11% | |
BlowingBubbles | -5.33% | -14.40% | -13.41% | -5.00% | -14.02% | -12.75% | |
RaceHorses | -9.20% | -25.79% | -27.27% | -9.24% | -25.53% | -24.35% | |
Average | -8.77% | -20.00% | -21.63% | -8.03% | -20.29% | -20.24% | |
Average | -7.34% | -17.21% | -16.65% | -7.95% | -17.91% | -17.44% |
IV-A Training Details
IV-A1 Training Data
We adopt the Vimeo-90k triplet dataset[41] for training. The Vimeo-90k dataset consists of 73,171 3-frame sequences with a fixed resolution of . During inference, the inputs for STENet consist of two reconstructed frames containing compression artifacts, while the outputs consist of an intermediate synthesized frame and two enhanced frames with suppressed artifacts. Compression artifacts are introduced into the training data to simulate real-world inference scenarios. Specifically, we compress the first and third frames in each triplet using the all intra (AI) configuration, while employing three consecutive original frames as ground-truth frames. The quantization parameter (QP) for each triplet is randomly selected from 22, 27, 32, 37, and 42.
IV-A2 Training Strategy
The STENet is trained using PyTorch on four Nvidia Geforce RTX 3060 GPUs. We use a Charbonnier penalty function[42] as the loss function and supervise all three output frames of STENet simultaneously:
(17) |
where is the output frames of STENet and denotes ground-truth frames. is set to as a constant. The loss function is optimized through the Adam optimizer with and . Before training, the STENet loads the pre-trained weights of IFRNet. The initial learning rates of IFRNet and other parts are set to and respectively and decay to with one cosine annealing[43]. For data augmentation, two input frames are randomly cropped into patches, and their order is randomly swapped. The batch size is 48, and the training lasts 300 epochs.
IV-B Experimental Results and Analysis
Class | Bitrate Reduction | Y-PSNR Improvement |
---|---|---|
Class A1 | -2.77% | 0.19% |
Class A2 | -4.04% | 0.18% |
Class B | -3.28% | 0.24% |
Class C | -5.29% | 0.31% |
Class D | -4.90% | 0.48% |
Average | -4.08% | 0.29% |
Class | Computational Time (CPU Only) | |
---|---|---|
Encoding | Decoding | |
Class A1 | 274% | 270045% |
Class A2 | 256% | 247660% |
Class B | 302% | 335021% |
Class C | 197% | 208032% |
Class D | 190% | 180358% |
Average | 272% | 261266% |
The proposed method is integrated into the VVC reference software VTM-15.0. The performance of our method is evaluated using VTM-15.0 as the benchmark anchor under RA configuration, with the compression performance measured by the Bjøntegaard Delta rate (BD-rate)[44]. A negative (positive) BD-rate percentage value indicates an increase (decrease) in compression efficiency performance. Follow the JVET NNVC CTC[40], each sequence is encoded with five QP, including 22, 27, 32, 37, and 42. Table I reports the PSNR-based BD-Rate and the MSIM-based BD-Rate for each sequence. As can be observed, the average PSNR-based BD-Rate and average MSIM-based BD-Rate are -7.34%/-17.21%/-16.65% and -7.95%/-17.91%/-17.44% respectively.
The visual comparisons of the reconstructed frames between the anchor VTM-15.0 and our proposed method are depicted in Fig. 4. We use two high QP values, 37 and 42, to better observe visual differences. It can be observed that the anchor fails to handle the ME and MC of moving objects, such as the basketball’s edge and the woman’s back. Additionally, noticeable blocking and ringing artifacts persist in the reconstructed frames of the anchor, for instance, in the horse’s tail. In contrast, the reconstructed frames produced by our method exhibit higher quality. For example, the basketball edge appears smoother, the woman’s back is distinct, and the horse’s tail is more continuous. Our method effectively alleviates the issues present in the anchor. Specifically, RFS generates virtual reference frames to provide more references for the current to-be-coded frame, while PFE filters the reconstructed frames to eliminate artifacts.



Original



Original Crop



VTM-15.0 - QP37



Ours - QP37



VTM-15.0 - QP42



Ours - QP42
The Rate-Distortion (RD) curves for four distinct sequences at various resolutions are illustrated in Fig.5. Each point on the curve depicts the distortion level introduced by the encoder to the coded video signal at a target bitrate. The red curves depict the results of our proposed method, while the black curves represent the results of VTM-15.0. As depicted in Fig.5, the red points are located to the top left of the black ones, indicating that our approach achieves lower bitrates while having higher image quality. This perspective is proved by Table II, which highlights the average bitrate reduction and average Y-PSNR improvement achieved by the proposed method compared to VTM-15.0. Our method demonstrates a notable 4.08% average bitrate reduction. Regarding video quality enhancement, Table II reveals an average Y-PSNR gain of approximately 0.29%.
Each to-be-coded CU searches for matching reference blocks in two reference frames during bi-prediction. The more similarity between to-be-coded CUs and reference blocks, the fewer bitrates are required for encoding the residuals. Fig.6 illustrates the reference block selection for each CU under RA configuration. Rectangles of different colors represent different reference block usages of to-be-coded CUs. Red rectangles indicate that both reference blocks for the to-be-coded CU are from synthesized frames, blue rectangles signify that only one reference block is from a synthesized frame, and black rectangles denote that neither of the two reference blocks is synthesized. As shown in Fig.6, many to-be-coded CUs choose synthesized blocks as references, especially at lower quality levels (higher QP). It is evident that synthesized frames provide good references for to-be-coded frames besides those original reconstructed frames.




The complexity analyses of our proposed method are provided in Table III. The time complexity is measured by evaluating the computational time ratio between our method and VTM-15.0. Compared to VTM-15.0, our method exhibits an average encoding complexity increase of 272%. It should be noted that all times are evaluated using a CPU. The time increment primarily arises from the network inference of RFS, PFE, and JISE, as well as the rate-distortion optimization (RDO) processes for the synthesized reference frames. On the decoder side, only the network inference requires additional time. Regarding network complexity, the number of parameters is 9.8M, and the number of multiply-accumulates per pixel is 3792K.
IV-C Ablation Study
Class | BD-rate | ||||||||
---|---|---|---|---|---|---|---|---|---|
(a) Without RFS | (b) Without PFE | (c) Without Joint Training | |||||||
Y-PSNR | U-PSNR | V-PSNR | Y-PSNR | U-PSNR | V-PSNR | Y-PSNR | U-PSNR | V-PSNR | |
Class A1 | 5.00% | 8.37% | 8.89% | 1.43% | 4.03% | 4.88% | -0.08% | 1.16% | 0.11% |
Class A2 | 6.77% | 11.60% | 8.97% | 0.96% | 3.97% | 2.31% | 0.72% | 1.31% | 1.41% |
Class B | 5.26% | 10.62% | 9.62% | 1.27% | 4.73% | 3.21% | 0.66% | 1.76% | 0.70% |
Class C | 7.26% | 15.23% | 15.24% | 1.36% | 6.53% | 6.94% | 0.45% | 1.93% | 1.56% |
Class D | 8.02% | 15.74% | 16.67% | 1.25% | 6.47% | 6.78% | 0.36% | 2.27% | 1.65% |
Average | 6.46% | 12.47% | 12.07% | 1.26% | 5.24% | 4.87% | 0.44% | 1.74% | 1.10% |
Class | Runtime | |||
---|---|---|---|---|
(a) Without RFS | (b) Without PFE | |||
Encoding | Decoding | Encoding | Decoding | |
Class A1 | 61% | 33% | 91% | 85% |
Class A2 | 63% | 33% | 91% | 85% |
Class B | 58% | 32% | 89% | 84% |
Class C | 70% | 33% | 92% | 84% |
Class D | 70% | 33% | 92% | 85% |
Average | 66% | 33% | 91% | 85% |
Class | BD-rate | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Hu et al. [10] | Jia et al. [11] | Ours (Without PFE) | Ours | |||||||||
Y-PSNR | U-PSNR | V-PSNR | Y-PSNR | U-PSNR | V-PSNR | Y-PSNR | U-PSNR | V-PSNR | Y-PSNR | U-PSNR | V-PSNR | |
Class A1 | -0.74% | -1.85% | -2.35% | -3.30% | -7.14% | -7.48% | -4.15% | -8.09% | -8.82% | -5.22% | -10.81% | -12.14% |
Class A2 | -0.70% | -3.23% | -2.52% | -4.92% | -10.73% | -8.66% | -6.20% | -11.40% | -9.37% | -7.14% | -14.02% | -10.94% |
Class B | -0.81% | -2.88% | -3.33% | -3.50% | -8.97% | -8.70% | -4.56% | -10.80% | -9.91% | -5.72% | -14.08% | -12.03% |
Class C | -2.13% | -4.76% | -4.85% | -4.92% | -11.75% | -12.02% | -6.08% | -13.32% | -13.44% | -7.36% | -17.68% | -18.36% |
Class D | -3.08% | -5.24% | -6.90% | -6.84% | -12.69% | -14.12% | -7.34% | -13.75% | -15.20% | -8.55% | -18.37% | -20.20% |
Average | -1.54% | -3.67% | -4.12% | -4.69% | -10.32% | -10.34% | -5.66% | -11.62% | -11.51% | -6.80% | -15.22% | -14.93% |
The ablation study in this paper employs a “tool-off” test procedure. Specifically, the “tool-off” test for a given tool involves comparing the VTM software configured to disable that particular tool with the regular VTM software (where the tool would be enabled)[45]. The experimental results are quantified using the PSNR-based BD-Rate. If a tool brings the compression improvement, disabling it will result in a positive BD-Rate value. In the “tool-off” test, we use the proposed method as the anchor and individually disable RFS, PFE, and the joint training strategy of RFS and PFE to examine their respective effects on the proposed method. The performance comparison of the “tool-off” test is illustrated in Table IV, and the runtime comparison is shown in Table V.
IV-C1 Reference Frame Synthesis
RFS generates a virtual reference frame resembling the to-be-coded frame, effectively reducing residuals. We conduct a “tool-off” test to assess the significance of RFS, where RFS is disabled, and the proposed method serves as the anchor. The performance comparison, detailed in Table IV(a), reveals performance losses of 6.46%/12.47%/12.07% across three components when RFS is disabled. Simultaneously, as indicated in Table V(a), encoding and decoding times are reduced to 66% and 33%, respectively. RFS is the primary driver of performance improvement in our methodology and constitutes a substantial portion of the total runtime.
IV-C2 Post-processing Filter Enhancement
PFE operates out of the encoding and decoding loops, impacting only the final reconstructed frames. PFE will alleviate artifacts introduced by block partitioning and quantization. To evaluate the impact of PFE, we turn it off during inference, with the proposed method serving as the anchor. As depicted in Table IV(b), the absence of PFE leads to a noticeable decline in the performance of 1.26%/5.24%/4.87%. The more significant performance drop in chroma components implies that PFE is particularly effective in mitigating artifacts within chroma components. As illustrated in Table V(b), encoding time and decoding time are reduced to 91% and 85%, respectively. This indicates that PFE does not significantly contribute to the overall complexity of our approach.




IV-C3 Joint Training Strategy
As described in Section IV-A2, the synthesis and enhancement pipelines are jointly trained within a single STENet, significantly simplifying the training cost. To investigate the impact of joint training on performance, we separately trained two STENets. For the first STENet, we only supervise the synthesized frame (second frame) and maintain all training settings unchanged. For the second STENet, we supervise the enhanced frames (first and third frames) and keep the original training settings.
Regarding network inference, the first STENet’s synthesis pipeline is utilized in RFS, while the second STENet’s enhancement pipeline is utilized in PFE. As for JISE, the single STENet is replaced by a cascade of two pipelines. As shown in Table IV, training two separate STENets for RFS and PFE resulted in a performance decrease of 0.44%/1.74%/1.10%. It indicates that the joint training strategy not only simplifies training costs but also enhances performance in the proposed method.
IV-D Comparison with Existing Methods
We compare the proposed method with two other RFS approaches [10] and [11]. Hu et al. [10] introduce the Error-Corrected Auto-Corrected Network (ECAR-Net) for inter prediction in video coding. Jia et al. [11] present a deep reference frame generation method, incorporating an interpolation network to enhance bi-directional prediction. Both methods are integrated into VTM-15.0 and operated under RA configuration. Experimental results are detailed in Table VI, with results for [10] and [11] obtained from their respective papers. We maintain the same QP values to align their experimental results as [10] only provides QP values for 22, 27, 32, and 37. For a fair comparison, we report the performance of the proposed method with PFE disabled. Evidently, with PFE disabled, our method outperforms theirs in each class. When PFE is enabled, our method exhibits even more significant performance gains. On average, a coding efficiency enhancement of up to 6.80% is achieved in the Y component, surpassing the 1.54% and 4.69% improvements reported in [10] and [11], respectively. The proposed method’s performance in the U and V components far exceeds that of [10] and [11].
V Conclusions
This paper presents joint RFS and PFE for VVC. We propose a well-designed STENet, in which two input frames are enhanced through its enhancement pipeline and utilized to synthesize an intermediate frame through its synthesis pipeline. During RFS, two reconstructed frames are sent into STENet’s synthesis pipeline to synthesize a virtual reference frame, which is inserted into two RPLs as additional reference frames. During PFE, two reconstructed frames are sent into STENet’s enhancement pipeline to alleviate their artifacts and distortions. In addition, we propose JISE to reduce inference time. Meanwhile, the proposed method only needs one joint training, which saves lots of resources. Integrated into the VVC reference software VTM-15.0, the proposed method could achieve -7.34%/-17.21%/-16.65% PSNR-based BD-rate on average for three components under RA configuration. Experimental results demonstrate the effectiveness and superiority of our proposed method over existing methods.
References
- [1] B. Bross, Y.-K. Wang, Y. Ye, S. Liu, J. Chen, G. J. Sullivan, and J.-R. Ohm, “Overview of the versatile video coding (VVC) standard and its applications,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 10, pp. 3736–3764, 2021.
- [2] B. Bross, J. Chen, J.-R. Ohm, G. J. Sullivan, and Y.-K. Wang, “Developments in international video coding standardization after AVC, with an overview of versatile video coding (VVC),” Proceedings of the IEEE, vol. 109, no. 9, pp. 1463–1493, 2021.
- [3] D. Liu, Y. Li, J. Lin, H. Li, and F. Wu, “Deep learning-based video coding: A review and a case study,” ACM Computing Surveys, vol. 53, no. 1, pp. 11:1–11:35, 2021.
- [4] G. Lu, W. Ouyang, D. Xu, X. Zhang, C. Cai, and Z. Gao, “DVC: An end-to-end deep video compression framework,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 10998–11007, 2019.
- [5] J. Li, B. Li, and Y. Lu, “Neural video compression with diverse contexts,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 22616–22626, 2023.
- [6] J. Li, B. Li, and Y. Lu, “Neural video compression with feature modulation,” arXiv preprint arXiv:2402.17414, 2024.
- [7] W.-J. Chien, L. Zhang, M. Winken, X. Li, R.-L. Liao, H. Gao, C.-W. Hsu, H. Liu, and C.-C. Chen, “Motion vector coding and block merging in the versatile video coding standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 10, pp. 3848–3861, 2021.
- [8] L. Zhao, S. Wang, X. Zhang, S. Wang, S. Ma, and W. Gao, “Enhanced ctu-level inter prediction with deep frame rate up-conversion for high efficiency video coding,” in IEEE International Conference on Image Processing, pp. 206–210, 2018.
- [9] Y. Guo, Z. Liu, Z. Chen, and S. Liu, “Deep inter coding with interpolated reference frame for hierarchical coding structure,” in IEEE Visual Communications and Image Processing, pp. 302–305, 2020.
- [10] Y. Hu, W. Yang, J. Liu, and Z. Guo, “Deep inter prediction with error-corrected auto-regressive network for video coding,” ACM Transactions on Multimedia Computing, Communications and Applications, vol. 19, no. 1s, pp. 33:1–33:22, 2023.
- [11] J. Jia, Y. Zhang, H. Zhu, Z. Chen, Z. Liu, X. Xu, and S. Liu, “Deep reference frame generation method for VVC inter prediction enhancement,” IEEE Transactions on Circuits and Systems for Video Technology, pp. 1–1, 2023.
- [12] M. Karczewicz, N. Hu, J. Taquet, C.-Y. Chen, K. Misra, K. Andersson, P. Yin, T. Lu, E. François, and J. Chen, “VVC in-loop filters,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 10, pp. 3907–3925, 2021.
- [13] Y. Li, L. Zhang, and K. Zhang, “Convolutional neural network based in-loop filter for VVC intra coding,” in IEEE International Conference on Image Processing, pp. 2104–2108, 2021.
- [14] H. Wang, G. Ren, T. Ouyang, J. Zhang, W. Han, Z. Liu, and Z. Chen, “Perceptual in-loop filter for image and video compression,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 1769–1772, 2022.
- [15] F. Zhang, C. Feng, and D. R. Bull, “Enhancing VVC through cnn-based post-processing,” in IEEE International Conference on Multimedia and Expo, pp. 1–6, 2020.
- [16] D. Ma, F. Zhang, and D. R. Bull, “MFRNet: A new cnn architecture for post-processing and in-loop filtering,” IEEE Journal of Selected Topics in Signal Processing, vol. 15, no. 2, pp. 378–387, 2021.
- [17] M. Santamaria, Y.-H. Lam, F. Cricri, J. Lainema, R. G. Youvalari, H. Zhang, M. M. Hannuksela, E. Rahtu, and M. Gaubbuj, “Content-adaptive convolutional neural network post-processing filter,” in IEEE International Symposium on Multimedia, pp. 99–106, 2021.
- [18] H. Zhang, C. Jung, D. Zou, and M. Li, “WCDANN: A lightweight cnn post-processing filter for vvc-based video compression,” IEEE Access, vol. 11, pp. 83400–83413, 2023.
- [19] H. Sun, Z. Cheng, M. Takeuchi, and J. Katto, “Enhanced intra prediction for video coding by using multiple neural networks,” IEEE Transactions on Multimedia, vol. 22, no. 11, pp. 2764–2779, 2020.
- [20] T. Dumas, F. Galpin, and P. Bordes, “Iterative training of neural networks for intra prediction,” IEEE Transactions on Image Processing, vol. 30, pp. 697–711, 2021.
- [21] Z. Zhao, S. Wang, S. Wang, X. Zhang, S. Ma, and J. Yang, “Enhanced bi-prediction with convolutional neural network for high-efficiency video coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 29, no. 11, pp. 3291–3301, 2019.
- [22] Y. Li, D. Liu, H. Li, L. Li, F. Wu, H. Zhang, and H. Yang, “Convolutional neural network-based block up-sampling for intra frame coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 28, no. 9, pp. 2316–2330, 2018.
- [23] J. Lin, D. Liu, H. Yang, H. Li, and F. Wu, “Convolutional neural network-based block up-sampling for HEVC,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 29, no. 12, pp. 3701–3715, 2019.
- [24] S. Ma, X. Zhang, C. Jia, Z. Zhao, S. Wang, and S. Wang, “Image and video compression with neural networks: A review,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 30, no. 6, pp. 1683–1698, 2020.
- [25] E. Alshina, S. Liu, A. Segall, F. Galpin, J. Li, R.-L. Liao, D. Rusanovskyy, T. Shao, M. Wien, and P. Wu, “JVET AHG report: Neural network-based video coding,” in document JVET-AD0011, ITU-T/ISO/IEC Joint Video Experts Team, Apr. 2023.
- [26] Y. Li, J. Li, C. Lin, K. Zhang, L. Zhang, F. Galpin, T. Dumas, H. Wang, M. Coban, J. Ström, et al., “Designs and implementations in neural network-based video coding,” arXiv preprint arXiv:2309.05846, 2023.
- [27] D. Ma, F. Zhang, and D. R. Bull, “BVI-DVC: A training database for deep video compression,” IEEE Transactions on Multimedia, vol. 24, pp. 3847–3858, 2022.
- [28] X. Xu, S. Liu, and Z. Li, “Tencent video dataset (TVD): A video dataset for learning-based visual data compression and analysis,” arXiv preprint arXiv:2105.05961, 2021.
- [29] F. Galpin, T. Dumas, P. Bordes, P. Nikitin, F. L. Léannec, and E. F. and, “AHG11: Small ad-hoc deep-learning library,” in document JVET-W0181, ITU-T/ISO/IEC Joint Video Experts Team, Jul. 2021.
- [30] J. Liang, J. Cao, G. Sun, K. Zhang, L. Van Gool, and R. Timofte, “SwinIR: Image restoration using swin transformer,” in IEEE International Conference on Computer Vision Workshops, pp. 1833–1844, 2021.
- [31] S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, and M. Yang, “Restormer: Efficient transformer for high-resolution image restoration,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5718–5729, 2022.
- [32] M. Tassano, J. Delon, and T. Veit, “FastDVDnet: Towards real-time deep video denoising without flow estimation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1351–1360, 2020.
- [33] S. Huo, D. Liu, B. Li, S. Ma, F. Wu, and W. Gao, “Deep network-based frame extrapolation with reference frame alignment,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 3, pp. 1178–1192, 2021.
- [34] W. Bao, X. Chen, J. Jia, Y. Zhang, Z. Chen, Z. Liu, X. Xu, and S. Liu, “EE1-2.1: Deep reference frame generation for inter prediction enhancement,” in document JVET-AF0208, ITU-T/ISO/IEC Joint Video Experts Team, Oct. 2023.
- [35] Y. Zhang, H. Wang, H. Zhu, and Z. Chen, “Optical flow reusing for high-efficiency space-time video super resolution,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 33, no. 5, pp. 2116–2128, 2023.
- [36] L. Kong, B. Jiang, D. Luo, W. Chu, X. Huang, Y. Tai, C. Wang, and J. Yang, “IFRNet: Intermediate feature refine network for efficient frame interpolation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1959–1968, 2022.
- [37] K. C. Chan, S. Zhou, X. Xu, and C. C. Loy, “On the generalization of BasicVSR++ to video deblurring and denoising,” arXiv preprint arXiv:2204.05308, 2022.
- [38] W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, and Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874–1883, 2016.
- [39] F. Bossen, X. Li, K. Sharman, V. Seregin, and K. Suehring, “VTM and HM common test conditions and software reference configurations for SDR 4:2:0 10-bit video,” in document JVET-Y2010, ITU-T/ISO/IEC Joint Video Experts Team, Jan. 2022.
- [40] E. Alshina, R.-L. Liao, S. Liu, and A. Segall, “JVET common test conditions and evaluation procedures for neural network-based video coding technology,” in document JVET-AE2016, ITU-T/ISO/IEC Joint Video Experts Team, Jul. 2023.
- [41] T. Xue, B. Chen, J. Wu, D. Wei, and W. T. Freeman, “Video enhancement with task-oriented flow,” International Journal of Computer Vision, vol. 127, no. 8, pp. 1106–1125, 2019.
- [42] W.-S. Lai, J.-B. Huang, N. Ahuja, and M.-H. Yang, “Deep laplacian pyramid networks for fast and accurate super-resolution,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5835–5843, 2017.
- [43] I. Loshchilov and F. Hutter, “SGDR: stochastic gradient descent with warm restarts,” in International Conference on Learning Representations, 2017.
- [44] G. Bjontegaard, “Improvements of the bd-psnr model,” in document ITU-T SG16 Q.6 document VCEG-AI11, Jul. 2008.
- [45] J. Pfaff, A. Filippov, S. Liu, X. Zhao, J. Chen, S. De-Luxán-Hernández, T. Wiegand, V. Rufitskiy, A. K. Ramasubramonian, and G. Van der Auwera, “Intra prediction and mode coding in VVC,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 10, pp. 3834–3847, 2021.