22institutetext: Kyung Hee University
22email: [email protected]
IntereStyle: Encoding an Interest Region for Robust StyleGAN Inversion
Abstract
Recently, manipulation of real-world images has been highly elaborated along with the development of Generative Adversarial Networks (GANs) and corresponding encoders, which embed real-world images into the latent space. However, designing encoders of GAN still remains a challenging task due to the trade-off between distortion and perception. In this paper, we point out that the existing encoders try to lower the distortion not only on the interest region, e.g., human facial region but also on the uninterest region, e.g., background patterns and obstacles. However, most uninterest regions in real-world images are located at out-of-distribution (OOD), which are infeasible to be ideally reconstructed by generative models. Moreover, we empirically find that the uninterest region overlapped with the interest region can mangle the original feature of the interest region, e.g., a microphone overlapped with a facial region is inverted into the white beard. As a result, lowering the distortion of the whole image while maintaining the perceptual quality is very challenging. To overcome this trade-off, we propose a simple yet effective encoder training scheme, coined IntereStyle, which facilitates encoding by focusing on the interest region. IntereStyle steers the encoder to disentangle the encodings of the interest and uninterest regions. To this end, we filter the information of the uninterest region iteratively to regulate the negative impact of the uninterest region. We demonstrate that IntereStyle achieves both lower distortion and higher perceptual quality compared to the existing state-of-the-art encoders. Especially, our model robustly conserves features of the original images, which shows the robust image editing and style mixing results. We will release our code with the pre-trained model after the review.
Keywords:
StyleGAN, robust GAN inversion, interest region, interest disentanglement, uninterest filter1 Introduction

Recently, as Generative Adversarial Networks (GANs) [15] have been remarkably developed, real-world image editing through latent manipulation has been prevalent [27, 28, 33, 30, 32, 25]. Especially, the strong disentangled property of StyleGAN [18, 19] latent space, i.e., , enables scrupulous image editing [32, 22], which can change only desirable features, e.g., facial expression, while maintaining the others, e.g., identity and hairstyle. For editing the image precisely with StyleGAN, it is required to get the suitable style latent, from which StyleGAN can reconstruct the image that has low distortion, high perceptual quality, and editability without deforming the feature of the original image.
Though StyleGAN is generally known to construct the image with high perceptual quality, the original style space is not enough to represent every real-world image with low distortion. Consequently, a vast majority of recent StyleGAN encoders, including optimization-based methods, embed images into space [1, 24, 3]. uses the identical style vector for every layer in StyleGAN, obtained by the mapping function. On the other hand, space provides a different style vector per layer and can even provide a random style vector in . However, as the distribution of style latent is far from , reconstructed images show low perceptual quality and editability [30, 25]. Consequently, lowering the distortion while keeping the other factors is still challenging.

In this paper, we claim that training to lower distortion on the entire region of the image directly is undesirable. In most cases, images contain regions that cannot be generated due to the inherent limitation of generators. Figure 1 shows clear examples of real-world images in the facial domain, which contain regions that are infeasible to be generated, e.g., hats, accessories, and noisy backgrounds. Therefore, an encoder needs to concentrate on the generable region for inversion while ignoring the un-generable region (e.g., non-facial region for StyleGAN trained with FFHQ). This strategy helps the latents inverted from the generable region to be close to , which show high perceptual quality and editability, as shown in Figure 1.
Another observation is that an attempt to reconstruct the region which is not generable induces severe distortion even on the other generable regions. For example, in Figure 2, a hand overlapped with the facial region is not generable by GAN encoders. Restyle [3], which shows the lowest distortion among all encoder-based inversion models until now, tries to lower distortion on the hand too, which rather causes catastrophic distortions on the nose and chin.
In the light of these observations, it is important to distinguish the region to be reconstructed elaborately from the rest. Here we define the term interest region, where the model focuses on the precise reconstruction with low distortion and high perceptual quality. Practically, in most cases, the interest region is aligned with the generable region of the image. For example, in facial image generation, the main interests are the face and hair parts of the output images, which are easier to generate than backgrounds. By focusing on the interest region, we can reduce distortion without any additional task, such as an attempt to encode latent excessively far from [8].
1.0.1 Contributions
We propose a simple yet effective method for training a StyleGAN encoder, coined IntereStyle, which steers the encoder to invert given images by focusing on the interest region. In particular, we introduce two novel training schemes for the StyleGAN encoder: (a) Interest Disentanglement (InD) and (b) Uninterest Filter (UnF). First, InD precludes the style latent, which includes the information on the uninterest region, from distorting the inversion result of the interest region. Second, UnF filters the information of the uninterest region, which prevents our model from redundantly attending to the uninterest region. UnF boosts the effect of InD by forcing the model not to focus on the uninterest region overly. In addition, we propose a very simple yet effective scheme for determining the interest region, required only at the training stage.
We demonstrate that IntereStyle, combined with the iterative refinement [3], effectively reduces the distortion at the interest region of CelebA-HQ-test dataset. To the best of our knowledge, IntereStyle achieves the lowest distortion among the existing state-of-the-art encoder-based StyleGAN inversion models without generator tuning. Moreover, we qualitatively show that our model robustly preserves features of the original images even with overlapped obstacles, while other baselines fail to. Lastly, we show the experimental results for image editing via InterFaceGAN [28], StyleCLIP [22], and style mixing [18] results, where our model shows remarkably robust outputs when input images contain significant noises, e.g., obstacles on the face.
2 Related Work
2.0.1 GAN Inversion
GAN inversion aims to transform given real-world images into latent vectors from which a pre-trained GAN model can reconstruct the original image. In the early stage of GAN inversion, the majority of models rely partially [36, 4, 5, 35] or entirely [11, 23, 1, 2, 16, 10, 12, 26] on the optimization steps per image. Though the optimization-based models show high inversion qualities, these models should perform numerous optimization steps per every input image [20], which are extremely time-consuming. Thus, training encoders that map images into the latent space has been prevalent to invert images in the real-time domain [29, 35, 24, 30, 3, 31]. However, regardless of encoding methods, the existing state-of-the-art GAN inversion models focus on the whole region of images [3, 24, 30, 2], including both interest and uninterest regions. We propose that focusing mainly on the interest region during GAN inversion improves the perceptual quality and editability of inverted images.


2.0.2 GAN Inversion Trade-Off
The desirable GAN inversion should consider both distortion and perceptual quality of inverted images [6, 30, 25]. However, due to the trade-off between two features, maintaining low distortion while enhancing the perceptual quality remains a challenging task [30, 25]. Especially in StyleGAN, an inverted image from the latent far from distribution achieves lower distortion [1, 2, 24] but shows lower perceptual quality [30] than the image from distribution. Moreover, the latent far from distribution shows lower editability [30, 25], which makes editing the inverted image harder. Here the variance among latents for all layers can be used as an indicator of distance from , where shows a zero variance due to the identical latent per layer.111Technically, we should identify whether the latents are from the mapping network of StyleGAN or not, but for simplicity, we only use the variance. As shown in Figure 3, the existing StyleGAN inversion models that show low distortion but suffer from low perceptual quality, e.g., [24] and Restyle [3], show relatively high variance among latents for all layers of StyleGAN. Especially, Figure 3(b) shows that Restyle gradually increases the variance of latents as the iteration refinement progresses. In the case of [30], it encodes images into latents close to but with high distortion. In contrast to the existing methods, our model focuses on lowering distortion at the interest region, i.e., hair and face. Since it is much easier than lowering at the uninterest region, i.e., irregular backgrounds, hats, and accessories, our model successfully achieves lower distortion than the existing models while avoiding the drop of high perceptual quality.
3 Method


In this section, we propose a simple yet effective StyleGAN encoder training scheme named IntereStyle. We first introduce our notation and the model architecture. Then, we introduce how to determine the interest region in input images. Next, we propose two novel methods: interest disentanglement and uninterest filter in Section 3.3 and Section 3.4, respectively. Finally, we describe the whole training scheme of IntereStyle in Section 3.5.
3.1 Notation and Architecture
Our architecture is shown in Figure 4, which is based on [24] model, combined with the iterative refinement [3, 31]. At -th iteration of the iterative refinement, our encoder receives a latent calculated at the previous step, 222When , we utilize as an average latent of StyleGAN., together with a pair of images. The pair consists of , where is a decoded result of via generator , i.e., , and is a preprocessed ground truth image, , by our proposed method in Section 3.4. targets to encode the difference between and into the latent, . Consequently, can yield an image , which is more similar to than , by decoding the latent . Our model iteratively refines the latent with a total of iterations. Finally, we utilize a loss function for training, consisting of the weighted sum of , [34], and [24]. We explain the details of each loss in Appendix A.
3.2 Interest Region
To guide the model to focus on the interest region, we need to label the interest region first. The interest region can be designated arbitrarily according to the usage of the inversion model. For instance, facial and hair regions for the facial domain, and the whole body for the animal domain can be set as the interest region. For labeling this interest region, the off-the-shelf segmentation networks are used, which is described in Section 4. However, directly using the segmentation masks from networks causes the distortion of facial boundaries in the generated image. To accommodate the boundary information, we use the dilated segmentation mask containing the interest region, as shown in Figure 5. Without dilation, the loss term on the interest region cannot penalize the inverted face on the uninterest region. Consequently, we dilate the mask to penalize the overflowed reconstruction of the interest region boundary. Though the small part of the uninterest region would be included in the interest region through the dilation, we empirically find that our model still precisely generates the interest region without any distortion of boundaries. We visually show the effect of mask dilation at the ablation study in Section 4.1.1.
3.3 Interest Disentanglement
To enforce the model to invert the interest region into the latent space precisely, we should train the model to concentrate on the interest region regardless of the uninterest region. However, due to the spatial-agnostic feature of Adaptive Instance Normalization (AdaIN) [21], inverted style latents considering the uninterest region may deteriorate the inversion quality of the interest region. To prevent the encoded style of the uninterest region from deforming the inverted result of the interest region, the inversion of each region should be disentangled.
As encodes the difference of the input pair of images in ideal, the latent obtained by encoding the pair of images that only differ on the uninterest region does not contain the information of the interest region. In other words, the decoding results from the latents w and should be the same on the interest region. Motivated by the above, we propose a simple method named Interest Disentanglement (InD) to distinguish the inversions of the interest region and uninterest region. We construct the pair of input images for InD as follows: the original image , and the same but multiplied with the interest region mask. Then, as shown in Figure 4, we can yield the pair of images which only differs in the uninterest region, and the corresponding latent from , . Ideally, the information in is solely related to the uninterest region, which implies generates the interest region robustly even is added. Consequently, we define InD loss, as follows;
(1) |
where is the inversion result from the latent . We apply Interest Disentanglement only at the training stage, which enables the inference without the interest region mask. We empirically find that IntereStyle focuses on the interest region without any prior mask given, after the training.
3.4 Uninterest Filter
At the early steps of the iterative refinement, focuses on reducing the distortion of the uninterest region [3]. Due to the spatial-agnostic feature of AdaIN, we claim that excessively focusing on the uninterest region hinders the inversion of the interest region. We propose a method named Uninterest Filter (UnF), to make concentrate on the interest region at every iteration consistently. UnF eliminates details of the uninterest region, which is inherently infeasible to reconstruct. Thus, can reduce the redundant attention on the uninterest region for the low distortion. In detail, UnF eases calculating by blurring the uninterest region of at each iteration, with a low-pass Gaussian Filter with radius , . As shown in Figure 4, UnF gradually reduces the radius of Gaussian filter of as iterations progress, with the following two reasons; First, the redundant attention on the uninterest region is considerably severe at the early stage of the iterations [3]. Consequently, we should blur the image at the early iterations heavily. Second, excessive blur results in the severe texture difference between the interest and uninterest region. We claim that is implicitly trained to encode the difference of the whole region, which can be biased to produce the blurred region when the blurred images are consistently given. For the realistic generation, the input at the -th iteration, is deblurred, i.e., identical to . We calculate the input image at the -th iteration as below:
(2) |
3.5 Training IntereStyle
At the training stage, we jointly train the model with the off-the-shelf encoder training loss [24] and . However, in contrast to Restyle [3], which back-propagates times per batch, we back-propagate only once after the -th iteration is over. Thus, ours show relatively faster training speed compared to Restyle. Our final training loss is defined as below:
(3) |
Our proposed methods, InD and UnF are synergistic at the training; While InD disentangles the inversion of the uninterest region, UnF forces to look at the interest region. Though applying to the images multiplied with inherently drives to focus on the interest region, InD is essential for robust training. Without , we find implicitly contains information of the uninterest region into , which affects the inversion of the interest region by AdaIN.
4 Experiments
In this section, we briefly introduce our datasets and baselines first. The implementation details are described in Appendix A. Then, we compare the inversion results with baselines and ablation scenarios, both in qualitative and quantitative ways. Next, we compare the image manipulation of our model, together with baselines. Finally, we look into the iterative scheme of our model with Restyle. Though we mainly show the results on the facial domain, we note that our method shows remarkable results in various domains. We show the experimental results on the animal domain in Figure 6 briefly and the plenty experimental results in Appendix D.
4.0.1 Datasets
For the facial domain, we trained the encoder using the FFHQ dataset, consisting of 70,000 human facial images. For the validation, we used the CelebA-HQ test set, consisting of 2,800 human facial images. We did not add or change any alignment or augmentation procedures compared to the existing encoder training methods [24, 3, 30] for a fair comparison. To generate the interest and uninterest region masks, we used the pre-trained Graphonomy [14] model. For the animal domain, we used AFHQ wild dataset [9] for training and validation, which consists of 4,730 and 500 images, respectively. We used the pre-trained DEtection TRansformer (DE-TR; [7]) for obtaining the interest region.
4.0.2 Baselines
We compared our model with the several well-known StyleGAN encoders: IDGI [35], [24], [30], and Restyle [3]. Moreover, in the case of the qualitative comparison of inversion, we additionally compared it with the optimization-based model [1, 2], which is well-known for its outstanding performance. For the baseline models, we used the pre-trained weights that are publicly available for evaluation. Please refer to Appendix C for more detailed information of each baseline.
4.1 Inversion Evaluation
4.1.1 Qualitative Evaluation

Model | IDGI | pSp | e4e | IntereStyle | ||||||||
Region | Interest | Face | Interest | Face | Interest | Face | Interest | Face | Interest | Face | Interest | Face |
0.030 | 0.010 | 0.018 | 0.006 | 0.023 | 0.007 | 0.015 | 0.004 | 0.021 | 0.007 | 0.013 | 0.003 | |
LPIPS | 0.116 | 0.053 | 0.095 | 0.046 | 0.111 | 0.051 | 0.088 | 0.040 | 0.109 | 0.054 | 0.075 | 0.036 |
ID Similarity | 0.18 | 0.56 | 0.47 | 0.65 | 0.51 | 0.68 |
Figure 6 shows the inverted images of IntereStyle and two StyleGAN inversion baselines, Image2StyleGAN [1] and Restyle [3]. In this figure, we show the entire inversion results along with the cropped images, which correspond to the areas of overlapping of the interest and uninterest regions. Without our robust inversion schemes, mere attempts to lower distortion over the entire region often occurred severe artifacts or feature deformations. Indeed, the baselines produced artifacts, which severely lower the perceptual quality. In addition, they mangled features of original images in some cases. For instance, turned the microphone into a white beard, which does not exist in the original image. The optimization-based inversion relatively mitigated artifacts among the baselines, but still suffered from them. Moreover, it required more than 200 times longer inference time compared to the one of IntereStyle. In contrast, IntereStyle showed the most robust outputs compared to all baselines, which best preserved the features of original images without artifacts.
In Figure 7, we qualitatively showed the effectiveness of the mask dilation. Without dilation, the model could not precisely reconstruct the original boundary of the interest region, which was denoted as the red region. In contrast, with the mask dilation, our model minimized the boundary deformation.

Method | LPIPS | |||
Interest | Face | Interest | Face | |
Baseline Restyle [3] | 0.015 | 0.004 | 0.088 | 0.040 |
+ on the interest region | 0.013 | 0.005 | 0.084 | 0.038 |
+ Interest Disentanglement (InD) | 0.012 | 0.003 | 0.078 | 0.037 |
+ Uninterest Filter (UnF) | 0.013 | 0.003 | 0.075 | 0.036 |
4.1.2 Quantitative Evaluation
We used and LPIPS [34] losses and measured ID similarity [24] by utilizing the pre-trained Curricularface [17], which shows the state-of-the-art performance on facial recognition. We measured the quality on the interest region, the facial and hair regions in this paper, which need to be inverted precisely. To this end, we multiplied the interest region mask at the calculation of ID similarity to preclude the facial recognition model from being affected by the non-facial region. Since the facial recognition performance is dependent on the features of the non-facial region [13], the inverted images are prone to be identified as similar faces with the original images due to the resemblance of non-facial regions. To compare the models solely on the facial region, we should wipe out the uninterest region.
As shown in Table 1, IntereStyle showed low distortion on both the interest and facial regions and preserved the identity well simultaneously. We conclude that focusing on the interest region is indeed helpful for robust inversion. Table 2 shows the ablation study by sequentially applying each component of our method to measure the effectiveness of the model performance. InD reduced the negative effect of the uninterest region, which indeed lowered distortion, compared to naïvely applying on the interest region. UnF lowered LPIPS by forcing the model to preserve features of the interest region. Please refer to Appendix D for more detailed results of each ablation model.


4.2 Editing via Latent Manipulation
Inversion of GAN is deeply related to the image manipulation on the latent space. In this section, we compare the quality of edited images produced by various StyleGAN inversion models [24, 30, 3], manipulated via InterFaceGAN [28] and StyleCLIP [22] methods, and style mixing [18, 19]. Figure 8 shows the results of editing real-world images via InterFaceGAN and StyleCLIP, together with the inversion results. We changed three attributes for each method; smile, age, and pose for InterFaceGAN, and “smile”, “lipstick”, and “Mohawk hairstyle” for StyleCLIP. Our model showed high inversion and perceptual qualities consistently among various editing scenarios, even with strong makeups or obstacles. However, and missed important features of images, such as makeups or eye shapes. Moreover, produced artifacts in several editing scenarios. In the cases of and , they failed to robustly handle obstacles. In the right side of Figure 8(a), produced severe artifacts around the mouth, while totally changed the shape. Moreover, the Restyle-based models showed low editability in specific cases, such as “Mohawk”.

To attribute to the superior disentanglement feature of StyleGAN latent space [28], we can separately manipulate the coarse and fine features of images. Following the settings from the StyleGAN [18] experiment, we took styles corresponding to either coarse, middle, or fine spatial resolution, respectively, from the latent of source B, and the others were taken from the latent of source A. Moreover, we mixed more than one hard case, e.g., obstacles on faces and extreme poses, to evaluate the robustness of our model. As shown in Figure 9, our model showed higher perceptual quality on the interpolated images compared to . produced images with texture shift, i.e., images with cartoon-like texture, distorted facial shape, and undesirable artifacts during the style mixing. In contrast, our model generated stable facial outputs. Additional qualitative results are shown in Appendix D.2.


4.3 Iterative Refinement of IntereStyle
We compared the progress of iterative refinement between Restyle [3] and IntereStyle in Figure 10. Restyle reconstructed most of the coarse features within a few steps, while the variance of Restyle increased consistently as iteration progressed. In other words, the reduction of distortion is marginal, though Restyle excessively focuses on this. Consequently, the latent from Restyle was located far from , which yields an image with low perceptual quality. In contrast, IntereStyle concentrated on the interest region that could be generated without a broad extension from . Consequently, IntereStyle effectively reduced distortion on the interest region by iteration while maintaining high perceptual quality.
5 Conclusions
For StyleGAN inversion, focusing on the interest region is essential but under-explored yet. We found excessive attention on the uninterest region occurs the drop of perceptual quality and high distortion on the interest region. We proposed a simple yet effective StyleGAN encoder training scheme, coined IntereStyle, composed of Interest Disentanglement and Uninterest Filter. We demonstrated that IntereStyle showed both low distortion and high perceptual inversion quality, and enabled various latent manipulations robustly for image editing. We look forward to our work to be widely used in future research or the industry field, which needs a delicate inversion of the interest region for image editing.
6 Acknowledgement
This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (No. 2021R1G1A1094379), and in part by the Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korea Government (MSIT) (Artificial Intelligence Innovation Hub) under Grant 2021-0-02068.
References
- [1] Abdal, R., Qin, Y., Wonka, P.: Image2stylegan: How to embed images into the stylegan latent space? In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 4432–4441 (2019)
- [2] Abdal, R., Qin, Y., Wonka, P.: Image2stylegan++: How to edit the embedded images? In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 8296–8305 (2020)
- [3] Alaluf, Y., Patashnik, O., Cohen-Or, D.: Restyle: A residual-based stylegan encoder via iterative refinement. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 6711–6720 (2021)
- [4] Bau, D., Strobelt, H., Peebles, W., Wulff, J., Zhou, B., Zhu, J.Y., Torralba, A.: Semantic photo manipulation with a generative image prior. arXiv preprint arXiv:2005.07727 (2020)
- [5] Bau, D., Zhu, J.Y., Wulff, J., Peebles, W., Strobelt, H., Zhou, B., Torralba, A.: Seeing what a gan cannot generate. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 4502–4511 (2019)
- [6] Blau, Y., Michaeli, T.: The perception-distortion tradeoff. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 6228–6237 (2018)
- [7] Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision. pp. 213–229. Springer (2020)
- [8] Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European conference on computer vision (ECCV). pp. 801–818 (2018)
- [9] Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: Stargan v2: Diverse image synthesis for multiple domains. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 8188–8197 (2020)
- [10] Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: Uncovering the local semantics of gans. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5771–5780 (2020)
- [11] Creswell, A., Bharath, A.A.: Inverting the generator of a generative adversarial network. IEEE transactions on neural networks and learning systems 30(7), 1967–1974 (2018)
- [12] Daras, G., Odena, A., Zhang, H., Dimakis, A.G.: Your local gan: Designing two dimensional local attention mechanisms for generative models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 14531–14539 (2020)
- [13] Deng, J., Guo, J., Xue, N., Zafeiriou, S.: Arcface: Additive angular margin loss for deep face recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 4690–4699 (2019)
- [14] Gong, K., Gao, Y., Liang, X., Shen, X., Wang, M., Lin, L.: Graphonomy: Universal human parsing via graph transfer learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7450–7459 (2019)
- [15] Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Advances in neural information processing systems 27 (2014)
- [16] Gu, J., Shen, Y., Zhou, B.: Image processing using multi-code gan prior. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 3012–3021 (2020)
- [17] Huang, Y., Wang, Y., Tai, Y., Liu, X., Shen, P., Li, S., Li, J., Huang, F.: Curricularface: adaptive curriculum learning loss for deep face recognition. In: proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 5901–5910 (2020)
- [18] Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 4401–4410 (2019)
- [19] Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 8110–8119 (2020)
- [20] Kim, H., Choi, Y., Kim, J., Yoo, S., Uh, Y.: Stylemapgan: Exploiting spatial dimensions of latent in gan for real-time image editing. arXiv preprint arXiv:2104.14754 (2021)
- [21] Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2337–2346 (2019)
- [22] Patashnik, O., Wu, Z., Shechtman, E., Cohen-Or, D., Lischinski, D.: Styleclip: Text-driven manipulation of stylegan imagery. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 2085–2094 (2021)
- [23] Raj, A., Li, Y., Bresler, Y.: Gan-based projector for faster recovery with convergence guarantees in linear inverse problems. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 5602–5611 (2019)
- [24] Richardson, E., Alaluf, Y., Patashnik, O., Nitzan, Y., Azar, Y., Shapiro, S., Cohen-Or, D.: Encoding in style: a stylegan encoder for image-to-image translation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2287–2296 (2021)
- [25] Roich, D., Mokady, R., Bermano, A.H., Cohen-Or, D.: Pivotal tuning for latent-based editing of real images. arXiv preprint arXiv:2106.05744 (2021)
- [26] Saha, R., Duke, B., Shkurti, F., Taylor, G.W., Aarabi, P.: Loho: Latent optimization of hairstyles via orthogonalization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1984–1993 (2021)
- [27] Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of gans for semantic face editing. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 9243–9252 (2020)
- [28] Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: Interpreting the disentangled face representation learned by gans. IEEE transactions on pattern analysis and machine intelligence (2020)
- [29] Tewari, A., Elgharib, M., Bharaj, G., Bernard, F., Seidel, H.P., Pérez, P., Zollhofer, M., Theobalt, C.: Stylerig: Rigging stylegan for 3d control over portrait images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6142–6151 (2020)
- [30] Tov, O., Alaluf, Y., Nitzan, Y., Patashnik, O., Cohen-Or, D.: Designing an encoder for stylegan image manipulation. ACM Transactions on Graphics (TOG) 40(4), 1–14 (2021)
- [31] Wei, T., Chen, D., Zhou, W., Liao, J., Zhang, W., Yuan, L., Hua, G., Yu, N.: A simple baseline for stylegan inversion. arXiv preprint arXiv:2104.07661 (2021)
- [32] Wu, Z., Lischinski, D., Shechtman, E.: Stylespace analysis: Disentangled controls for stylegan image generation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 12863–12872 (2021)
- [33] Yang, G., Fei, N., Ding, M., Liu, G., Lu, Z., Xiang, T.: L2m-gan: Learning to manipulate latent space semantics for facial attribute editing. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2951–2960 (2021)
- [34] Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 586–595 (2018)
- [35] Zhu, J., Shen, Y., Zhao, D., Zhou, B.: In-domain gan inversion for real image editing. In: European conference on computer vision. pp. 592–608. Springer (2020)
- [36] Zhu, J.Y., Krähenbühl, P., Shechtman, E., Efros, A.A.: Generative visual manipulation on the natural image manifold. In: European conference on computer vision. pp. 597–613. Springer (2016)