This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Beyond GFVC: A Progressive Face Video Compression Framework with Adaptive Visual Tokens

Bolin Chen    Shanzhi Yin    Zihan Zhang    Jie Chen    Ru-Ling Liao   
Lingyu Zhu
   Shiqi Wang and Yan Ye
City University of Hong Kong DAMO Academy, Alibaba Group
Abstract

Recently, deep generative models have greatly advanced the progress of face video coding towards promising rate-distortion performance and diverse application functionalities. Beyond traditional hybrid video coding paradigms, Generative Face Video Compression (GFVC) relying on the strong capabilities of deep generative models and the philosophy of early Model-Based Coding (MBC) can facilitate the compact representation and realistic reconstruction of visual face signal, thus achieving ultra-low bitrate face video communication. However, these GFVC algorithms are sometimes faced with unstable reconstruction quality and limited bitrate ranges. To address these problems, this paper proposes a novel Progressive Face Video Compression framework, namely PFVC, that utilizes adaptive visual tokens to realize exceptional trade-offs between reconstruction robustness and bandwidth intelligence. In particular, the encoder of the proposed PFVC projects the high-dimensional face signal into adaptive visual tokens in a progressive manner, whilst the decoder can further reconstruct these adaptive visual tokens for motion estimation and signal synthesis with different granularity levels. Experimental results demonstrate that the proposed PFVC framework can achieve better coding flexibility and superior rate-distortion performance in comparison with the latest Versatile Video Coding (VVC) codec and the state-of-the-art GFVC algorithms. The project page can be found at https://github.com/Berlin0610/PFVC.

1 Introduction

In recent years, visual data compression has been greatly powered by Artificial Intelligence Generated Content (AIGC) techniques, where the compression paradigms have been shifted from signal level [1, 2] to generative representation [3, 4, 5] or even semantic information [6, 7]. Different from traditional hybrid-based coding frameworks, these generative video coding solutions have shown advantageous Rate-Distortion (RD) performance for visual content with strong priors and can support more diverse scenarios like metaverse-related or user-specified animation applications. In particular, these existing generative codecs mainly follow the early philosophy of Model-Based Coding (MBC) [8, 9] techniques and explore compact spatial/temporal representations to bridge the video compression & generation tasks.

Taking face videos as an example, they exhibit strong prior knowledge and statistical regularities, which can be easily incorporated into the Generative Face Video Compression (GFVC) frameworks [3] towards high-efficiency visual communication. Different from early MBC techniques, these existing GFVC methods are mainly rooted in the latest face animation/reenactment models [10, 11] and rely on the strong inference capabilities of deep models to achieve high-quality video reconstruction. More specifically, the GFVC encoder can characterize high-dimension visual signal into compact representations (like 2D keypoints [10, 12], 3D keypoints [13], compact feature [5] and facial semantics [7]) for economical transmission costs via the analysis model. Furthermore, the GFVC decoder is capable of transforming these decoded compact information into fine-grained motion fields and employing the deep generative model to reconstruct realistic face signal. Although compact representations can facilitate promising RD performance within the GFVC pipelines, they also sometimes suffer unstable reconstruction quality and cover limited bitrate ranges, greatly limiting broader applications of GFVC. In other words, the compact representation technique is like a double-edged sword, which can greatly optimize signal representation costs but sometimes cause irreversible losses in signal reconstruction. In addition, this technique can only produce fixed-dimension representation and lack abilities to adapt to network bandwidth, leading to poor coding flexibility.

In view of these existing limitations, this paper proposes a novel Progressive Face Video Compression framework based upon adaptive visual tokens, namely PFVC, aiming at facilitating generative face video communication towards bandwidth intelligence and robust reconstruction. In particular, inspired by the recent advancements of visual tokenization for the efficient signal synthesis [14], the proposed PFVC framework transforms complex visual face signal into different-granularities visual motion tokens in a progressive manner. As such, these visual tokens are capable of hierarchically estimating facial motions and reconstructing the corresponding face contents according to different bandwidth scenarios. Moreover, the proposed progressive token strategy can allow these GFVC pipelines to be no longer limited to relying on compact representations for signal reconstruction, thus enhancing the effectiveness and robustness of the face generation process. The contributions of this paper are summarized as follows,

  • We propose a novel PFVC framework that could progressively characterize high-dimensional visual face signal into adaptive sparse tokens and actualize different-quality face signal reconstruction. As such, the proposed PFVC scheme enjoys the advantages of coding flexibility, bandwidth intelligence and reliable reconstruction for face video communication.

  • We design an adaptive visual tokenization mechanism to describe facial motions at different granularities, such that the coded bitstream can be featured with different visual tokens in a progressive manner. Moreover, these adaptive visual tokens can be further transformed into pixel-wise dense motion fields for the temporal evolution inference via the implicit motion learning and GAN-based frame generation schemes.

  • We develop a progressive training strategy for generative video coding model, where different-granularities visual tokens are progressively introduced in each training stage and also guided with probability learning. Therefore, the proposed PFVC only needs one model to be compatible with overall bandwidth coverage instead of using multiple models.

Refer to caption

Figure 1: Overview of the proposed PFVC framework towards bandwidth intelligence.

2 The Proposed PFVC Framework

In this section, we give an overall introduction regarding the proposed PFVC encoding/decoding processes. In addition, we discuss the design principle of each important component/module in the PFVC framework. Finally, we describe a new model training strategy for progressive video coding.

2.1 Overall Encoding/Decoding Processes

As shown in Figure 1, the proposed PFVC framework is mainly grounded on the adaptive visual tokens, thus exploring exceptional capabilities in the trade-offs between reconstruction robustness and bandwidth intelligence. At the encoder side, the input video signal can be classified as the key-reference frame KK (e.g., the first frame of this video) and subsequent inter frames Il(1ln,lZ)I_{l}\left(1\leq l\leq n,l\in Z\right). The key-reference frame KK is intra-coded by an off-the-shelf VVC encoder [2], thus providing the rich texture reference for the subsequent frame reconstruction. Afterwards, the VVC-reconstructed key-reference frame K^\hat{K} and these remaining inter frames IlI_{l} are all characterized into a series of adaptive visual tokens in a progressive manner. Then, the network bandwidth scenario will further determine which granularity of visual tokens can be compressed. Then, these determined tokens are coded using inter prediction, quantization and entropy coding, where a context-adaptive arithmetic coder [15, 16] is employed to achieve optimal compression performance.

The proposed PFVC decoder aims at using the decoded adaptive visual tokens to reconstruct face video at different quality levels. In particular, the off-the-shelf VVC decoder is employed to reconstruct the key-reference frame K^\hat{K}. In addition, the adaptive tokens of inter frames can be obtained by context-based entropy decoding and difference compensation. These decoded visual tokens are developed to infer the temporal motion evolution and transformed into pixel-wise dense motion fields, where their motion quality can be gradually improved with better token granularity. Finally, the powerful generative model can facilitate to reconstruct the face video with the guidance of the decoded key-reference frames and these inferred dense motion fields.

Thanks to the adaptive visual tokens, the proposed PFVC framework enjoys several benefits that are not available to other existing GFVC algorithms. First, our proposed PFVC can characterize high-dimensional visual face signal into compact tokens with different granularities, ensuring greater feature adaptability in a learnable manner. In addition, different from the existing GFVC algorithms, which can only control the RD trade-offs by adjusting Quantization Parameter (QP) levels of the key-reference frame, our proposed PFVC can also achieve coding flexibility and bandwidth intelligence by transmitting visual tokens of different granularities. Furthermore, finer-granularity tokens can improve the face reconstruction robustness and stability, facilitating to solve the inaccurate motion estimation and unreliable generation problems that sometimes plague the existing GFVC pipelines due to compact representations. Finally, our proposed PFVC only needs one model to output visual features of different granularities and achieve face reconstruction of different qualities instead of using multiple GFVC models, which greatly expands the application possibilities.

2.2 Key Techniques Introduction

Herein, we will introduce all key techniques in the proposed PFVC framework, including adaptive visual token representation, implicit motion field evolution and GAN-based signal reconstruction. These designed schemes have greatly empowered the inference capability of the PFVC towards promising RD performance and bandwidth intelligence.

Adaptive Sparse Tokens Representation. These high-dimensional visual signal K^\hat{K} or IlI_{l} can be transformed to adaptive sparse tokens with different granularities via a specifically-designed token extractor. In particular, these face frames first execute down-sampling operation ϕ()\phi\left(\cdot\right) through a scale factor ss, and then fed into the feature network f()\mathbb{N}_{f}\left(\cdot\right) that is mainly composed of a U-Net architecture and a series of convolution/normalization operations to obtain motion features motion\mathcal{F}_{motion} with the size of 1×16×161\times 16\times 16. The motion features entailing accurate head posture and facial expressions can be learned in an unsupervised manner,

motion=f(ϕ(X,s)),\vspace{-0.7em}{\mathcal{F}_{motion}=\mathbb{N}_{f}\left(\phi\left(X,s\right)\right)}, (1)

where XX could represent the VVC-reconstructed key-reference frame K^\hat{K} or the inter frames IlI_{l}.

Afterwards, these learned motion features are further fed into a token network t()\mathbb{N}_{t}\left(\cdot\right) that is essentially an encoder-decoder translator with three Fully-Connected (FC) layers. Specifically, at the encoder part te()\mathbb{N}_{te}\left(\cdot\right) of this token network, the 2-D motion features are first flatten into a 1-D tokens with the dimension of 256, and then are sequentially sampled into three other 1-D tokens via FC layers with the dimensions of 144, 64 and 16. It should be noted that the number of FC layers and the specific dimension of each FC layer can be randomly set and herein we just provide an implementation example. The produced adaptive token list tokens\mathcal{L}_{tokens} is denoted as,

tokens=te(motion),\vspace{-0.7em}{\mathcal{L}_{tokens}=\mathbb{N}_{te}\left(\mathcal{F}_{motion}\right)}, (2)

where the dimension of adaptive tokens tokens\mathcal{L}_{tokens} is 256, 144, 64 and 16. In other words, depending on the given bandwidth environment, different-granularities visual tokens can be further coded into bitstream and actualize variable bitrate. After decoding these tokens ^tokens\mathcal{\hat{L}}_{tokens}, they are input into the decoder part td()\mathbb{N}_{td}\left(\cdot\right) of the token network for the corresponding reconstruction of motion features ^motion\mathcal{\hat{F}}_{motion},

^motion=td(^tokens).\vspace{-0.7em}{\mathcal{\hat{F}}_{motion}=\mathbb{N}_{td}\left(\mathcal{\hat{L}}_{tokens}\right)}. (3)

Overall, the proposed token network can realize feature sparsity and produce visual tokens of different granularities, thus providing great possibilities to explore compression scalability.

Implicit Motion Field Evolution. Given the decoded key-reference frame K^\hat{K}, the extracted motion features motionK^\mathcal{F}_{motion}^{\hat{K}} from K^\hat{K} and the decoded motion features ^motionIl\mathcal{\hat{F}}_{motion}^{I_{l}} from IlI_{l}, we design an implicit motion field evolution scheme to achieve effective motion transformations from feature to pixel-wise fields. Different from CFTE [17]/CTTR [5] that both use the sparse-to-dense motion estimation strategy to generate a coarse deformed frame and then transform this frame for motion fields, our proposed scheme can directly use finer-granularity feature to infer motion fields via neural network learning. In particular, an up-sampled learning network φ()\varphi\left(\cdot\right) is employed to perform scale transformation for motionK^\mathcal{F}_{motion}^{\hat{K}} and ^motionIl\mathcal{\hat{F}}_{motion}^{I_{l}}. In addition, difference between these scaled features is calculated,

DiffIl,K^=φ(^motionIl)φ(motionK^).\vspace{-0.6em}{Diff_{\left\langle I_{l},\hat{K}\right\rangle}=\varphi\left(\mathcal{\hat{F}}_{motion}^{I_{l}}\right)-\varphi\left(\mathcal{F}_{motion}^{\hat{K}}\right)}. (4)

As an implicit motion change information, DiffI,KDiff_{\left\langle I,K\right\rangle} is treated as the input together with the down-sampled K^\hat{K} into the motion prediction network m()\mathbb{N}_{m}\left(\cdot\right), aiming at estimating pixel-wise dense motion map denseIl\mathcal{M}^{I_{l}}_{dense} and occlusion map occlusionIl\mathcal{M}^{I_{l}}_{occlusion} for face signal reconstruction. The process can be formulated as follows,

denseIl=P1(m(concat(K^,DiffIl,K^))),\mathcal{M}^{I_{l}}_{dense}=P_{1}\left(\mathbb{N}_{m}\left(concat\left(\hat{K},Diff_{\left\langle I_{l},\hat{K}\right\rangle}\right)\right)\right), (5)
occlusionIl=P2(m(concat(K^,DiffIl,K^))),\mathcal{M}^{I_{l}}_{occlusion}=P_{2}\left(\mathbb{N}_{m}\left(concat\left(\hat{K},Diff_{\left\langle I_{l},\hat{K}\right\rangle}\right)\right)\right), (6)

where P1()P_{1}\left(\cdot\right) and P2()P_{2}\left(\cdot\right) represent two different predicted operation.

GAN-based Signal Reconstruction. Analogous to the existing GFVC generator designs, we also adopt the flow-warped GAN module GAN()\mathbb{N}_{GAN}\left(\cdot\right) to reconstruct realistic face signal I^l\hat{I}_{l}. In particular, K^\hat{K} is first warped with MdenselM^{l}_{dense} for motion guidance and then the transformed features are further marked with occlusionIl\mathcal{M}^{I_{l}}_{occlusion} for occlusion impainting,

Il^=GAN(occlusionIlfw(K^,denseIl)),{\hat{I_{l}}=\mathbb{N}_{GAN}\left(\mathcal{M}^{I_{l}}_{occlusion}\odot f_{w}\left(\hat{K},\mathcal{M}^{I_{l}}_{dense}\right)\right)}, (7)

where fwf_{w} and \odot describe the back-warping and the Hadamard product operations.

2.3 Progressive Model Training Strategy

Table 1: Hyper parameter settings of the proposed progressive model training strategy
Basic Settings Stages S Epoch N Granularity G Probability P
N=200 M=4 G=[16, 64, 144, 256] S1S_{1} 1\sim40 [256] [1.0]
S2S_{2} 41\sim80 [144, 256] [0.7, 0.3]
S3S_{3} 81\sim120 [64, 144, 256] [0.5, 0.3, 0.2]
S4S_{4} 121\sim160 [16, 64, 144, 256] [0.45, 0.25, 0.15, 0.15]
S5S_{5} 161\sim200 [16, 64, 144, 256] [0.25, 0.25, 0.25, 0.25]

To better facilitate the proposed PFVC to be compatible with different token granularities and achieve high-quality motion transformations within the same model, we propose a new progressive model training strategy with the granular probability guidance. In particular, assuming that the PFVC model needs to be trained for NN epochs and can process MM granularities G={G1,G2,,GM}G=\left\{G_{1},G_{2},\dots,G_{M}\right\}, the overall model training will be divided into S={S1,S2,,SMN+1},MN+1ZS=\left\{S_{1},S_{2},\dots,S_{\frac{M}{N+1}}\right\},\frac{M}{N+1}\in Z stages. In each stage, the PFVC model will be assigned how to learn different token granularities with the given probabilistic ratios P={P1,P2,,PM}P=\left\{P_{1},P_{2},\dots,P_{M}\right\}. It should be mentioned that in each stage, the given probabilistic ratios for all granularities should be summed to 1. More importantly, to prevent the overall model capacity from being directly reduced by the coarse-grained tokens, the proposed progressive strategy prioritizes training of fine granularity in the early training stages and progressively introduce coarse-grained tokens in the subsequent stages. Besides, during the S1SMN1N+1S_{1}\sim S_{\frac{M-N-1}{N+1}} stages, the maximum probability will be assigned to the corresponding newly-introduced token granularity to improve the PFVC model’s acceptance of this granularity. As for the SMN+1S_{\frac{M}{N+1}} stage, the probability of each granularity will be set at the same ratio for stabilizing model training.

In addition to the proposed progressive training strategy, the perceptual loss per\mathcal{L}_{per}, adversarial loss adv\mathcal{L}_{adv} and feature matching loss fea\mathcal{L}_{fea} are adopted to supervise all PFVC stages in an end-to-end manner. The overall training loss can be described as,

total=λperper+λadvadv+λfeafea,\begin{array}[]{c}\mathcal{L}_{total}=\lambda_{per}\mathcal{L}_{per}+\lambda_{adv}\mathcal{L}_{adv}+\lambda_{fea}\mathcal{L}_{fea},\end{array} (8)

where per\mathcal{L}_{per} and adv\mathcal{L}_{adv} are used to ensure the fidelity and naturalness of the reconstructed frames I^l\hat{I}_{l}, and fea\mathcal{L}_{fea} is introduced to stabilize the the GAN model training. In addition, the hyper-parameters are set as follows: λper=10\lambda_{per}=10, λadv=1\lambda_{adv}=1 and λfea=10\lambda_{fea}=10.

3 Experimental Results

This section first introduces the experimental settings, and then provides performance comparisons in terms of objective RD results and subjective quality.

Refer to caption
(a) Rate-DISTS
Refer to caption
(b) Rate-LPIPS
Refer to caption
(c) Rate-FVD
Refer to caption
(d) Rate-MANIQA
Refer to caption
(e) Rate-PSNR
Refer to caption
(f) Rate-SSIM
Figure 2: Overall RD performance comparisons with VVC [2], FOMM [10] and CFTE [17] in terms of DISTS, LPIPS, FVD, MANIQA, PSNR and SSIM.

3.1 Experimental Settings

Implementation Details. We implement our proposed PFVC framework with Pytorch libraries, and train this model with the popular VoxCeleb training dataset [18] on 2 NVIDIA TESLA A100 GPUs. During the model training phase, the Adam optimizer is used with the parameters β1\beta_{1} = 0.5 & β2\beta_{2}= 0.999, and the learning rate is set at 0.0002 for model convergence. In addition to the synchronized BatchNorm initialization and the data repeating strategy, we also employ the progressive training strategy as shown in Table 1.

Compared Algorithms. To demonstrate the effectiveness of the proposed PFVC, we compare the state-of-the-art video coding standard VVC [2] and two representative GFVC algorithm (CFTE [17] and FOMM [10]). In particular, the testing dataset adopted in [7] includes 50 talking face videos from the VoxCeleb testing dataset [18], where each sequence contains 250 frames with the resolution of 256×\times256. The specific compared algorithm settings are provided as follows,

  • VVC anchor: is configured with the Low-Delay-Bidirectional (LDB) mode in VTM reference software 22.2, where the Quantization Parameters (QPs) are set at 32/35/37/40/42/45/47/52 and the coded signal format is YUV 4:2:0.

  • CFTE/FOMM: follows experimental testings specified in the JVET GFVC AhG test conditions [19]. In particular, the key-reference frame is converted to the YUV 4:2:0 format and compressed via the VTM reference software 22.2 with the QPs of 2/12/22/32/42/52. The subsequent inter frames can be represented into compact parameters and then compressed via a context-adaptive arithmetic coder.

  • PFVC (Proposed): Its test settings are basically the same as CFTE/FOMM except the feature representation of inter frames. Specifically, the proposed PFVC can characterize the inter frames into 4 different token granularities with the dimension of 16/64/144/256. Herein, the convex hull is performed on 6×4=246\times 4=24 RD points for a wider bitrate range and the optimal performance.

Evaluation Measures. Herein, we adopt 4 learning-based quality measures and 2 traditional quality measures to evaluate the quality of reconstructed face video, including DISTS [20], LPIPS [21], FVD [22], MANIQA [23], PSNR [24] and SSIM [25]. These 6 adopted measures mainly involve perceptual-level, temporal consistency, no-reference GAN-based and pixel-level evaluations, where smaller scores of DISTS/LPIPS/FVD and higher scores of MANIQA/PSNR/SSIM indicate better perceived quality. Therefore, Figure 4 uses 1-DISTS, 1- LPIPS and 1-FVD in order to present more coherent RD curves. In addition to reconstruction quality, the coding bitrate (i.e., kbps) is also provided to further evaluate the compression performance.

3.2 Performance Comparisons

Rate-Distortion Performance. Figure 4 shows the RD performance of the VVC codec, FOMM, CFTE and our proposed PFVC scheme in terms of DISTS, LPIPS, FVD, MANIQA, PSNR and SSIM. It can be seen that the proposed PFVC can outperform the VVC codec for DISTS/LPIPS/FVD/MANIQA quality evaluations at overall bitrate ranges and for PSNR/SSIM quality evaluations at relatively low bit rate ranges. In addition, compared with the CFTE/FOMM anchors, the proposed PFVC can achieve the highest performance for all metrics and effectively improve the reconstruction quality with similar bit consumption.

As for pixel-level measurement, the proposed PFVC and other GFVC algorithms cannot achieve any advantages compared with the VVC codec. This is largely due to the fact that generative compression algorithms aim at reconstructing face signal in perceptual-level feature domain instead of pixel-level domain, which cannot adapt to the pixel-level evaluations (PSNR/SSIM) [26]. More importantly, by introducing finer token granularity, the proposed PFVC can gradually improve the GFVC ability in pixel-level reconstruction. Although the performance can be improved, the adopted token granularities are still insufficient for the motion information supplementation, resulting in the saturation of its pixel-level reconstruction quality after a certain rate point.

Refer to caption

Figure 3: Subjective comparisons of reconstruction quality using the proposed PFVC with 4 different token granularities. More visual examples can be found in project page.
Refer to caption
(a) Visual Example 1
Refer to caption
(b) Visual Example 2
Figure 4: Subjective comparisons of reconstruction quality using different compression algorithms like VVC [2], FOMM [10] and CFTE [17] and our proposed PFVC at 2 different bitrate levels. More visual examples can be found in project page.

Subjective Quality. Figure 3 provides visual examples to verify that the proposed PFVC framework can reduce some annoying visual artifacts or recover more missing details via different-granularities visual token representations. For examples, with the provision of finer token granularity, it is possible to infer the hand details that originally do not exist in the key-reference frame. In addition, the finer token granularity can stabilize the background generation and facilitate higher face fidelity such as accurate mouth movement and eye direction.

As illustrated in Figure 4, the proposed PFVC algorithm is able to reconstruct face frames with high signal fidelity and accurate facial motion. In particular, at very low bit rates, the VVC reconstructed results exhibit blocking artifacts, while the proposed PFVC still can reconstruct vivid face signal. Besides, the facial motions in reconstructed video quality using FOMM and CFTE are not accurate, while the proposed PFVC can infer more precise head posture and facial expression.

4 Conclusion

This paper proposes a novel progressive generative face video compression framework to actualize coding flexibility, bandwidth intelligence and reliable reconstruction for face video communication. In particular, adaptive visual tokens are leveraged to describe facial motion changes in a progressive manner, whilst the deep generative model is employed to learn implicit motion fields from these visual tokens and perform temporal inference for high quality face reconstruction. Experimental results demonstrate that compared with the latest VVC codec and other GFVC algorithms, our proposed PFVC can achieve more promising RD performance and wider overall bitrate as well as quality converge.

References

References

  • [1] G. Sullivan et al., “Overview of the High Efficiency Video Coding (HEVC) standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 22, no. 12, pp. 1649–1668, 2012.
  • [2] B. Bross et al., “Overview of the Versatile Video Coding (VVC) standard and its applications,” IEEE Trans. Circuits Syst. Video Technol., vol. 31, no. 10, pp. 3736–3764, 2021.
  • [3] B. Chen et al., “Generative face video coding techniques and standardization efforts: A review,” in Data Compression Conference, 2024, pp. 103–112.
  • [4] Z. Wang et al., “Dynamic multi-reference generative prediction for face video compression,” in Proc. IEEE Int. Conf. Image Process., 2022, pp. 896–900.
  • [5] B. Chen et al., “Compact temporal trajectory representation for talking face video compression,” IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 11, pp. 7009–7023, 2023.
  • [6] P. Zhang et al., “Rethinking semantic image compression: Scalable representation with cross-modality transfer,” IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 8, pp. 4441–4445, 2023.
  • [7] B. Chen et al., “Interactive face video coding: A generative compression framework,” arXiv preprint arXiv:2302.09919, 2023.
  • [8] W. Schreiber et al., “Synthetic highs—an experimental tv bandwidth reduction system,” Journal of the SMPTE, vol. 68, no. 8, pp. 525–537, 1959.
  • [9] R. Lopez et al., “Head pose computation for very low bit-rate video coding,” in International Conference on Computer Analysis of Images and Patterns, 1995, pp. 440–447.
  • [10] A. Siarohin et al., “First order motion model for image animation,” Adv. Neural Inf. Process. Syst., vol. 32, pp. 7137–7147, 2019.
  • [11] F.-T. Hong et al., “Depth-aware generative adversarial network for talking head video generation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2022.
  • [12] G. Konuko, G. Valenzise, and S. Lathuilière, “Ultra-low bitrate video conferencing using deep image animation,” in IEEE Int. Conf. Acoust. Speech Signal Process.   IEEE, 2021, pp. 4210–4214.
  • [13] T. Wang et al., “One-shot free-view neural talking-head synthesis for video conferencing,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2021, pp. 10 039–10 049.
  • [14] Q. Yu et al., “An image is worth 32 tokens for reconstruction and generation,” arXiv preprint arXiv:2406.07550, 2024.
  • [15] J. Teuhola, “A compression method for clustered bit-vectors,” Inf. Process. Lett., vol. 7, no. 6, pp. 308–311, 1978.
  • [16] J. Cleary and I. Witten, “Data compression using adaptive coding and partial string matching,” IEEE Trans. Commun., vol. 32, no. 4, pp. 396–402, 1984.
  • [17] B. Chen et al., “Beyond keypoint coding: Temporal evolution inference with compact feature representation for talking face video compression,” in Data Compression Conference, 2022, pp. 13–22.
  • [18] A. Nagrani, J. S. Chung, and A. Zisserman, “Voxceleb: a large-scale speaker identification dataset,” in INTERSPEECH, 2017.
  • [19] S. McCarthy and B. Chen, “Test conditions and evaluation procedures for generative face video coding,” The JVET of ISO/IEC SC 29 and ITU-T SG16, doc. no. JVET-AG2035, January 2024.
  • [20] K. Ding et al., “Image quality assessment: Unifying structure and texture similarity,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
  • [21] R. Zhang et al., “The unreasonable effectiveness of deep features as a perceptual metric,” in others, 2018.
  • [22] T. Unterthiner, S. van Steenkiste, K. Kurach, R. Marinier, M. Michalski, and S. Gelly, “Fvd: A new metric for video generation,” in 7th International Conference on Learning Representations workshop, 2019.
  • [23] S. Yang et al., “Maniqa: Multi-dimension attention network for no-reference image quality assessment,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2022, pp. 1191–1200.
  • [24] W. Zhou et al., “Mean squared error: Love it or leave it? a new look at signal fidelity measures,” IEEE Signal Process. Mag., vol. 26, no. 1, pp. 98–117, 2009.
  • [25] Z. Wang et al., “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process., vol. 13, no. 4, pp. 600–612, 2004.
  • [26] Y. Li et al., “Perceptual quality assessment of face video compression: A benchmark and an effective method,” IEEE Trans. Multimedia, pp. 1–13, 2024.