This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Energy-Based Generative Cooperative Saliency Prediction

Jing Zhang1, Jianwen Xie2, Zilong Zheng3, Nick Barnes1
Abstract

Conventional saliency prediction models typically learn a deterministic mapping from an image to its saliency map, and thus fail to explain the subjective nature of human attention. In this paper, to model the uncertainty of visual saliency, we study the saliency prediction problem from the perspective of generative models by learning a conditional probability distribution over the saliency map given an input image, and treating the saliency prediction as a sampling process from the learned distribution. Specifically, we propose a generative cooperative saliency prediction framework, where a conditional latent variable model (LVM) and a conditional energy-based model (EBM) are jointly trained to predict salient objects in a cooperative manner. The LVM serves as a fast but coarse predictor to efficiently produce an initial saliency map, which is then refined by the iterative Langevin revision of the EBM that serves as a slow but fine predictor. Such a coarse-to-fine cooperative saliency prediction strategy offers the best of both worlds. Moreover, we propose a “cooperative learning while recovering” strategy and apply it to weakly supervised saliency prediction, where saliency annotations of training images are partially observed. Lastly, we find that the learned energy function in the EBM can serve as a refinement module that can refine the results of other pre-trained saliency prediction models. Experimental results show that our model can produce a set of diverse and plausible saliency maps of an image, and obtain state-of-the-art performance in both fully supervised and weakly supervised saliency prediction tasks.

1 Introduction

As a class-agnostic segmentation task, salient object detection has attracted a lot of attentions in the computer vision community for its close relationship to human visual perception. A salient region is a visually distinctive scene region that can be located rapidly and with little human effort. Salient object detection is commonly treated as a pixel-wise binary output of a deterministic prediction model in most recent works  (Wu, Su, and Huang 2019a; Qin et al. 2019; Wu, Su, and Huang 2019b; Wei, Wang, and Huang 2020; Wang et al. 2019; Xu et al. 2021). Despite the success of those recent models, the one-to-one deterministic mapping has prevented them from modeling the uncertainty of human salient object prediction, which is considered to be subjective (Itti, Koch, and Niebur 1998) and affected by biological factors (e.g., contrast sensitivity), contextual factors (e.g., task, experience, and interest), etc. In this way, it is more reasonable to represent visual saliency as a conditional probability distribution over a saliency map given an input image, and formulate the saliency prediction as a stochastic sampling process from the conditional distribution.

Generative models (Goodfellow et al. 2014; Kingma and Welling 2013; Xie et al. 2016, 2018a) have demonstrated their abilities to represent conditional distributions of high-dimensional data and produce multiple plausible outputs given the same input (Zhu et al. 2017; Xie et al. 2021b). In this work, we fit the saliency detection task into a generative framework, where the input image is the condition, and the goal is to generate multiple saliency maps, representing the “subjective nature” of human visual saliency. Zhang et al. (2020a, 2021) have used conditional variational auto-encoders (VAEs) (Kingma and Welling 2013; Sohn, Lee, and Yan 2015a), which are latent variable models (LVMs), to implicitly represent distributions of visual saliency. However, VAEs only learn a stochastic mapping from image domain to saliency domain, and lack an intrinsic cost function to evaluate the visual saliency output and guide the saliency prediction process. As to a prediction task, a cost function of solution is more reliable than a mapping function because the former is more generalizable than the latter.

In contrast, we propose to model the conditional distribution of visual saliency explicitly via an energy-based model (EBM) (Xie et al. 2016; Nijkamp et al. 2019), where the energy function defined on both image and saliency domains serves as a cost function of the saliency prediction. Given an input image, the saliency prediction can be achieved by performing sampling from the EBM via Markov chain Monte Carlo (MCMC) (Neal 2012) method, which is a gradient-based algorithm that searches local minima of the cost function of the EBM conditioned on the input image.

A typical high-dimensional EBM learns an energy function by MCMC-based maximum likelihood estimation (MLE), which commonly suffers from convergence difficulty and computational expensiveness of the MCMC process. Inspired by prior success of energy-based generative cooperative learning  (Xie et al. 2018a, 2021b), we propose the energy-based generative cooperative saliency prediction framework to tackle the saliency prediction task. Specifically, the framework consists of a conditional EBM whose energy function is parameterized by a bottom-top neural network, and a conditional LVM whose transformation function is parameterized by an encoder-decoder framework. The framework brings in an LVM as an ancestral sampler to initialize the MCMC computational process of the EBM for efficient sampling, so that the EBM can be learned efficiently. The EBM, in turn, refines the LVM’s generated samples via MCMC and feeds them back to the LVM, so that the LVM can learn its mapping function from the MCMC transition. Thus, the resulting cooperative saliency prediction process first generates an initial saliency map via a direct mapping and then refines the saliency map via an iterative process. This is a coarse-to-fine generative saliency detection, and corresponds to a fast-thinking and slow-thinking system (Xie et al. 2021b).

Moreover, based on the generative cooperative saliency prediction framework, we further propose a cooperative learning while cooperative recovering strategy for weakly supervised saliency learning, where each training image is associated with a partially observed annotation (e.g., scribble annotation (Zhang et al. 2020b)). At each learning iteration, the strategy has two sub-tasks: cooperative recovery and cooperative learning. As to the cooperative recovery sub-task, each incomplete saliency ground truth is firstly recovered in the low-dimensional latent space of the LVM via inference, and then refined by being pushed to the local mode of the cost landscape of the EBM via MCMC. For the cooperative learning sub-task, the recovered saliency maps are treated as pseudo labels to update the parameters of the framework as in the scenario of learning from complete data.

In experiments, we demonstrate that our framework can not only achieve state-of-the-art performances in both fully supervised and weakly supervised saliency predictions, but also generate diverse saliency maps from one input image, indicating the success of modeling the uncertainty of saliency prediction. Furthermore, we show that the learned energy function in the EBM can serve as a cost function, which is useful to refine the results from other pre-trained saliency prediction models.

Our contributions can be summarized as below:

  • We study generative modeling of saliency prediction, and formulate it as a sampling process from a probabilistic model using EBM and LVM respectively, which are new angles to model and solve saliency prediction.

  • We propose a generative cooperative saliency prediction framework, which jointly trains the LVM predictor and the EBM predictor in a cooperative learning scheme to offer reliable and efficient saliency prediction.

  • We generalize our generative framework to the weakly supervised saliency prediction scenario, in which only incomplete annotations are provided, by proposing the cooperative learning while recovering algorithm, where we train the model and simultaneously recover the unlabeled areas of the incomplete saliency maps.

  • We provide strong empirical results in both fully supevised and weakly supervised settings to verify the effectiveness of our framework for saliency prediction.

2 Related Work

We first briefly introduce existing fully supervised and weakly supervised saliency prediction models. We then review the family of generative cooperative models and other conditional deep generative frameworks.

Fully Supervised Saliency Prediction. Existing fully supervised saliency prediction models (Wang et al. 2018; Liu, Han, and Yang 2018; Wei, Wang, and Huang 2020; Liu et al. 2019; Qin et al. 2019; Wu, Su, and Huang 2019b, a; Wang et al. 2019, 2019; Wei et al. 2020; Xu et al. 2021) mainly focus on exploring image context information and generating structure-preserving predictions.Wu, Su, and Huang (2019b); Wang et al. (2019, 2019, 2018); Liu, Han, and Yang (2018); Liu et al. (2019); Wu, Su, and Huang (2019a); Xu et al. (2021) propose saliency prediction models by effectively integrating higher-level and lower-level features. Wei, Wang, and Huang (2020); Wei et al. (2020) propose an edge-aware loss term to penalize errors along object boundaries.Zhang et al. (2020a) present a stochastic RGB-D saliency detection network based on the conditional variational auto-encoder (Kingma and Welling 2013; Jimenez Rezende, Mohamed, and Wierstra 2014). In this paper, we introduce the conditional cooperative learning framework (Xie et al. 2018a, 2021b) to achieve probabilistic coarse-to-fine RGB saliency detection, where a coarse prediction is produced by a conditional latent variable model and then is refined by a conditional energy-based model. Our paper is the first work to use a deep energy-based generative framework for probabilistic saliency detection.

Weakly Supervised Saliency Prediction. Weakly supervised saliency prediction frameworks (Wang et al. 2017; Li, Xie, and Lin 2018a; Nguyen et al. 2019; Zhang et al. 2020b) attempt to learn predictive models from easy-to-obtain weak labels, including image-level labels (Wang et al. 2017; Li, Xie, and Lin 2018b), noisy labels (Nguyen et al. 2019; Zhang et al. 2018; Zhang, Han, and Zhang 2017) or scribble labels (Zhang et al. 2020b). In this paper, we also propose a cooperative learning while recovering strategy for weakly supervised saliency prediction, in which only scribble labels are provided and our model treats them as incomplete data and recovers them during learning.

Energy-Based Generative Cooperative Networks. Deep energy-based generative models (Xie et al. 2016), with energy functions parameterized by modern convolutional neural networks, are capable of modeling the probability density of high-dimensional data. They have been applied to image generation (Xie et al. 2016; Gao et al. 2018; Nijkamp et al. 2019; Du and Mordatch 2019; Grathwohl et al. 2020; Zhao, Xie, and Li 2021; Zheng, Xie, and Li 2021), video generation (Xie, Zhu, and Wu 2019), 3D volumetric shape generation (Xie et al. 2018b, 2020), and unordered point cloud generation (Xie et al. 2021a). The maximum likelihood learning of the energy-based model typically requires iterative MCMC sampling, which is computationally challenging. To relieve the computational burden of MCMC, the Generative Cooperative Networks (CoopNets) in Xie et al. (2018a) propose to learn a separate latent variable model (i.e. a generator) to serve as an efficient approximate sampler for training the energy-based model. Xie, Zheng, and Li (2021) propose a variant of CoopNets by replacing the generator with a variational auto-encoder (VAE) (Kingma and Welling 2014). Xie et al. (2021b) propose a conditional CoopNets for supervised image-to-image translation. Our paper proposes a conditional CoopNets for visual saliency prediction. Further, we generalize our model to the weakly supervised learning scenario by proposing a cooperative learning while recovering algorithm. In this way, we can learn from incomplete data for weakly supervised saliency prediction.

Conditional Deep Generative Models. Our framework belongs to the family of conditional generative models, which include conditional generative adversarial networks (CGANs) (Mirza and Osindero 2014) and conditional variational auto-encoders (CVAEs) (Sohn, Lee, and Yan 2015a). Different from existing CGANs (Luc et al. 2016; Zhang et al. 2018; Xue et al. 2017; Pan et al. 2017; Yu and Cai 2018; Hung et al. 2018; Souly, Spampinato, and Shah 2017), which train a conditional discriminator and a conditional generator in an adversarial manner, or CVAEs (Kohl et al. 2018; Zhang et al. 2020a, 2021), in which a conditional generator is trained with an approximate inference network, our model learns a conditional generator with a conditional energy-based model via MCMC teaching. Specifically, our model allows an additional refinement for the generator during prediction, which is lacking in both CGANs and CVAEs.

3 Cooperative Saliency Prediction

We will first present two types of generative modeling of saliecny prediction, i.e., the energy-based model (EBM) and the latent variable model (LVM). Then, we propose a novel generative saliency prediction framework, in which the EBM and the LVM are jointly trained in a generative cooperative manner, such that they can help each other for better saliency prediction in terms of computational efficiency and prediction accuracy. The latter aims to generate a coarse but fast prediction, and the former serves as a fine saliency predictor. The resulting model is a coarse-to-fine saliency prediction framework.

3.1 EBM as a Slow but Fine Predictor

Let XX be an image, and YY be its saliency map. The EBM defines a conditional distribution of YY given XX by:

pθ(Y|X)=pθ(Y,X)pθ(Y,X)𝑑Y=exp[Uθ(Y,X)]Z(X;θ),p_{\theta}(Y|X)=\frac{p_{\theta}(Y,X)}{\int p_{\theta}(Y,X)dY}=\frac{\exp[-U_{\theta}(Y,X)]}{Z(X;\theta)}, (1)

where the energy function Uθ(Y,X)U_{\theta}(Y,X), parameterized by a bottom-up neural network, maps the input image-saliency pair to a scalar, and θ\theta represent the network parameters. Z(X;θ)=exp[Uθ(Y,X)]𝑑YZ(X;\theta)=\int\exp[-U_{\theta}(Y,X)]dY is the normalizing constant. When UθU_{\theta} is learned and an image XX is given, the prediction of saliency YY can be achieved by Langevin sampling (Neal 2012), which makes use of the gradient of the energy function and iterates the following step:

Yτ+1=Yτδ22YUθ(Yτ,X)+δΔτ,Y_{\tau+1}=Y_{\tau}-\frac{\delta^{2}}{2}\frac{\partial}{\partial Y}U_{\theta}(Y_{\tau},X)+\delta\Delta_{\tau}, (2)

where τ\tau indexes the Langevin time step, δ\delta is the step size, and Δτ𝒩(0,ID)\Delta_{\tau}\sim\mathcal{N}(0,I_{D}) is a Gaussian noise term. The Langevin dynamics (Neal 2012) is initialized with Gaussian distribution and is equivalent to a stochastic gradient descent algorithm that seeks to find the minimum of the objective function defined by Uθ(Y,X)U_{\theta}(Y,X). The noise term Δτ\Delta_{\tau} is a Brownian motion that prevents gradient descent from being trapped by local minima of Uθ(Y,X)U_{\theta}(Y,X). The energy function Uθ(Y,X)U_{\theta}(Y,X) in Eq. (1) can be regarded as a trainable cost function of the task of saliency prediction. The prediction process via Langevin dynamics in Eq. (2) can be considered as finding YY to minimize the cost Uθ(Y,X)U_{\theta}(Y,X) given an input XX. Such a framework can learn a reliable and generalizable cost function for saliency prediction. However, due to the iterative sampling process in the prediction, EBM is slower than LVM, which adopts a mapping for direct sampling.

3.2 LVM as a Fast but Coarse Predictor

Let hh be a latent Gaussian noise vector. The LVM defines a mapping function Gα(X,h):[h,X]YG_{\alpha}(X,h):[h,X]\rightarrow Y that maps a latent vector h𝒩(0,Id)h\sim\mathcal{N}(0,I_{d}) together with an image XX to a saliency map YY. IdI_{d} is a dd-dimensional identity matrix. dd is the number of dimensionalities of hh. Specifically, the mapping function GG is parameterized by a noise-injected encoder-decoder network with skip connections and α\alpha contain all the learning parameters in the network. The LVM is given by:

Y=Gα(X,h)+ϵ,ϵ𝒩(0,σ2ID),Y=G_{\alpha}(X,h)+\epsilon,\epsilon\sim\mathcal{N}(0,\sigma^{2}I_{D}), (3)

where ϵ\epsilon is an observation residual and σ\sigma is a predefined standard deviation of ϵ\epsilon. The LVM in Eq. (3) defines an implicit conditional distribution of saliency YY given an image XX, i.e., pα(Y|X)=pα(Y|X,h)p(h)𝑑hp_{\alpha}(Y|X)=\int p_{\alpha}(Y|X,h)p(h)dh, where pα(Y|X,h)=𝒩(Gα(X,h),σ2ID)p_{\alpha}(Y|X,h)=\mathcal{N}(G_{\alpha}(X,h),\sigma^{2}I_{D}). The saliency prediction can be achieved by an ancestral sampling process that first samples a Gaussian white noise vector hh and then transforms it along with an image XX to a saliency map YY. Since the ancestral sampling is a direct mapping, it is faster than the iterative Langevin dynamics in the EBM. However, without a cost function as in the EBM, the learned mapping in the LVM is hard to be generalized to a new domain.

3.3 Cooperative Prediction with Two Predictors

We propose to predict image saliency by a cooperative sampling strategy. We first use the coarse saliency predictor (LVM) to generate an initial prediction Y^\hat{Y} via a non-iterative ancestral sampling, and then we use the fine saliency predictor (EBM) to refine the initial prediction via KK-step Langevin revision to obtain a revised saliency Y~\tilde{Y}. The process can be written as:

Y^=Gα(X,h^),h^𝒩(0,Id),{Y~0=Y^Y~τ+1=Y~τδ22Y~Uθ(Y~τ,X)+δΔτ\displaystyle\begin{aligned} &\hat{Y}=G_{\alpha}(X,\hat{h}),\hat{h}\sim\mathcal{N}(0,I_{d}),\\ &\begin{cases}\tilde{Y}_{0}=\hat{Y}&\\ \tilde{Y}_{\tau+1}=\tilde{Y}_{\tau}-\frac{\delta^{2}}{2}\frac{\partial}{\partial\tilde{Y}}U_{\theta}(\tilde{Y}_{\tau},X)+\delta\Delta_{\tau}\end{cases}\end{aligned} (4)

We call this process the cooperative sampling-based coarse-to-fine prediction. In this way, we take both advantages of these two saliency predictors in the sense that the fine saliency predictor (i.e., Langevin sampler) is initialized by the efficient coarse saliency predictor (i.e., ancestral sampler), while the coarse saliency predictor is refined by the accurate fine saliency predictor that aims to minimize a cost function UθU_{\theta}.

Since our conditional model represents a one-to-many mapping, the prediction is stochastic. To evaluate the learned model on saliency prediction tasks, we can draw multiple h^\hat{h} from the prior 𝒩(0,Id)\mathcal{N}(0,I_{d}) and use their average to generate Y^\hat{Y}, then a Langevin dynamics with the diffusion term being disabled (i.e., gradient descent) is performed to push Y^\hat{Y} to its nearest local minimum Y~\tilde{Y} based on the learned energy function. The resulting Y~\tilde{Y} is treated as a prediction of our model.

3.4 Cooperative Training of Two Predictors

We use the cooperative training method (Xie et al. 2018a, 2021b) to learn the parameters of the two predictors. At each iteration, we first generate synthetic examples via the cooperative sampling strategy shown in Eq. (4), and then the synthetic examples are used to compute the learning gradients to update both predictors. We present the update formula of each predictor below.

MCMC-based Maximum Likelihood Estimation (MLE) for the Fine Saliency Predictor. Given a training dataset {(Xi,Yi)}i=1n\{(X_{i},Y_{i})\}_{i=1}^{n}, we train the fine saliency predictor via MLE, which maximizes the log-likelihood of the data L(θ)=1ni=1nlogpθ(Yi|Xi)L(\theta)=\frac{1}{n}\sum_{i=1}^{n}\log p_{\theta}(Y_{i}|X_{i}), whose learning gradient is Δθ=1ni=1n{𝔼pθ(Y|Xi)[θUθ(Y,Xi)]θUθ(Yi,Xi)}\Delta{\theta}=\frac{1}{n}\sum_{i=1}^{n}\{\mathbb{E}_{p_{\theta}(Y|X_{i})}[\frac{\partial}{\partial\theta}U_{\theta}(Y,X_{i})]-\frac{\partial}{\partial\theta}U_{\theta}(Y_{i},X_{i})\}. We rely on the cooperative sampling in Eq. (4) to sample Yi~pθ(Y|Xi)\tilde{Y_{i}}\sim p_{\theta}(Y|X_{i}) to approximate the gradient:

Δθ1ni=1nθUθ(Y~i,Xi)1ni=1nθUθ(Yi,Xi).\Delta{\theta}\approx\frac{1}{n}\sum_{i=1}^{n}\frac{\partial}{\partial\theta}U_{\theta}(\tilde{Y}_{i},X_{i})-\frac{1}{n}\sum_{i=1}^{n}\frac{\partial}{\partial\theta}U_{\theta}(Y_{i},X_{i}). (5)

We can use Adam (Kingma and Ba 2015) with Δθ\Delta{\theta} to update θ\theta. We denote Δθ({Yi},{Y~i})\Delta{\theta}(\{Y_{i}\},\{\tilde{Y}_{i}\}) as a function of {Yi}\{Y_{i}\} and {Y~i}\{\tilde{Y}_{i}\}.

Maximum Likelihood Training of the Coarse Saliency Predictor by MCMC Teaching. Even though the fine saliency predictor learns from the training data, the coarse saliency predictor learns to catch up with the fine saliency predictor by treating {(X,Y~)}i=1n\{(X,\tilde{Y})\}_{i=1}^{n} as training examples. The learning objective is to maximize the log-likelihood of the samples drawn from pθ(Y|X)p_{\theta}(Y|X), i.e., L(α)=1ni=1nlogpα(Y~i|Xi)L(\alpha)=\frac{1}{n}\sum_{i=1}^{n}\log p_{\alpha}(\tilde{Y}_{i}|X_{i}), whose gradient can be computed by

Δα=i=1n𝔼hpα(h|Yi,Xi)[αlogpα(Yi,h|Xi)].\Delta{\alpha}=\sum_{i=1}^{n}\mathbb{E}_{h\sim p_{\alpha}(h|Y_{i},X_{i})}\left[\frac{\partial}{\partial\alpha}\log p_{\alpha}(Y_{i},h|X_{i})\right]. (6)

This leads to an MCMC-based solution that iterates (i) an inference step: inferring latent h~\tilde{h} by sampling from posterior distribution h~pα(h|Y,X)\tilde{h}\sim p_{\alpha}(h|Y,X) via Langevin dynamics, which iterates the following:

h~τ+1=h~τ+δ22h~logpα(Y,h~τ|X)+δΔτ,\displaystyle\tilde{h}_{\tau+1}=\tilde{h}_{\tau}+\frac{\delta^{2}}{2}\frac{\partial}{\partial\tilde{h}}\log p_{\alpha}(Y,\tilde{h}_{\tau}|X)+\delta\Delta_{\tau}, (7)

where Δτ𝒩(0,Id)\Delta_{\tau}\sim\mathcal{N}(0,I_{d}) and h~logpα(Y,h~|X)=1σ2(YGα(X,h~))h~Gα(X,h~)h~\frac{\partial}{\partial\tilde{h}}\log p_{\alpha}(Y,\tilde{h}|X)=\frac{1}{\sigma^{2}}(Y-G_{\alpha}(X,\tilde{h}))\frac{\partial}{\partial\tilde{h}}G_{\alpha}(X,\tilde{h})-\tilde{h}, and (ii) a learning step: with {h~i,Y~i,Xi}\{\tilde{h}_{i},\tilde{Y}_{i},X_{i}\}, we update α\alpha via Adam optimizer with

Δα1ni=1n1σ2(Y~iGα(Xi,h~i))αGα(Xi,h~i).\Delta{\alpha}\approx\frac{1}{n}\sum_{i=1}^{n}\frac{1}{\sigma^{2}}(\tilde{Y}_{i}-G_{\alpha}(X_{i},\tilde{h}_{i}))\frac{\partial}{\partial\alpha}G_{\alpha}(X_{i},\tilde{h}_{i}). (8)

Since GαG_{\alpha} is parameterized by a differentiable neural network, both hGα(X,h)\frac{\partial}{\partial h}G_{\alpha}(X,h) in Eq. (7) and αGα(Xi,h~i)\frac{\partial}{\partial\alpha}G_{\alpha}(X_{i},\tilde{h}_{i}) in Eq. (8) can be efficiently computed by back-propagation. We denote Δα({h~i},{Y~i})\Delta{\alpha}(\{\tilde{h}_{i}\},\{\tilde{Y}_{i}\}) as a function of {h~i}\{\tilde{h}_{i}\} and {Y~i}\{\tilde{Y}_{i}\}. Algorithm 1 presents a description of the cooperative learning algorithm of the fine and coarse saliency predictors.

Algorithm 1 Training the Cooperative Saliency Predictor

Input:
(1) Training images {Xi}in\{X_{i}\}_{i}^{n} with associated saliency maps {Yi}in\{Y_{i}\}_{i}^{n};
(2) maximal number of learning iterations TT.

Output: Parameters θ\theta and α\alpha

1:Initialize θ\theta and α\alpha with Gaussian noise
2:for t1t\leftarrow 1 to TT do
3:     Draw h^i𝒩(0,Id)\hat{h}_{i}\sim\mathcal{N}(0,I_{d})
4:     Sample initial prediction Y^i=Gα(Xi,h^i)\hat{Y}_{i}=G_{\alpha}(X_{i},\hat{h}_{i}).
5:     Revise Y^i\hat{Y}_{i} to obtain Y~i\tilde{Y}_{i} by Langevin dynamics in Eq. (2)
6:     Revise h^i\hat{h}_{i} to obtain h~i\tilde{h}_{i} by Langevin dynamics in Eq. (7)
7:     Update θ\theta with Δθ({Yi},{Y~i})\Delta\theta(\{Y_{i}\},\{\tilde{Y}_{i}\}) in Eq. (5) using Adam
8:     Update α\alpha with Δα({h~i},{Y~i})\Delta\alpha(\{\tilde{h}_{i}\},\{\tilde{Y}_{i}\}) in Eq. (8) using Adam
9:end for

4 Weakly Supervised Saliency Prediction

In Section 3, the framework is trained from fully-observed training data. In this section, we want to show that our generative framework can be modified to handle the scenario in which each image XiX_{i} only has a partial pixel-wise annotation YiY^{\prime}_{i}, e.g., scribble annotation (Zhang et al. 2020b). Since the saliency map for each training image is incomplete, directly applying the algorithm to the incomplete training data can lead to a failure of learning the distribution of saliency given an image. However, generative models are good at data recovery, therefore they can learn to recover the incomplete data. In our framework, we will leverage the recovery powers of both EBM and LVM to deal with the incomplete data in our cooperative learning algorithm, and this will lead to a novel weakly supervised saliency prediction framework.

To learn from incomplete data, our algorithm alternates the cooperative learning step and the cooperative recovery step. Both steps need a cooperation between EBM and LVM. The cooperative learning step is the same as the one used for fully observed data, except that it treats the recovered saliency maps, which are generated from the cooperative recovery step, as training data in each iteration. The following is the cooperative recovery step, which consists of two sub-steps driven by the LVM and the EBM respectively:

(i) Recovery by LVM in Latent Space. Given an image XiX_{i} and its incomplete saliency map YiY^{\prime}_{i}, the recovery of the missing part of YiY^{\prime}_{i} can be achieved by first inferring the latent vector hih^{\prime}_{i} based on the partially observed saliency information via hipα(h|Yi,Xi)h^{\prime}_{i}\sim p_{\alpha}(h|Y^{\prime}_{i},X_{i}), and then generating Y^i=Gα(Xi,hi)\hat{Y}^{\prime}_{i}=G_{\alpha}(X_{i},h^{\prime}_{i}) with the inferred latent vector hih^{\prime}_{i}. Let OiO_{i} be a binary mask, with the same size as YY^{\prime}, indicating the locations of visible annotations in YiY^{\prime}_{i}. OiO_{i} varies for different YiY^{\prime}_{i} and can be extracted from YiY^{\prime}_{i}. The Langevin dynamics for recovery iterates the same step in Eq. (7) except that hlogpα(Y,hτ|X)=1σ2(O(YGα(X,hτ)))hGα(X,hτ)hτ\frac{\partial}{\partial h}\log p_{\alpha}(Y^{\prime},h_{\tau}|X)=\frac{1}{\sigma^{2}}(O\circ(Y-G_{\alpha}(X,h_{\tau})))\frac{\partial}{\partial h}G_{\alpha}(X,h_{\tau})-h_{\tau}, where \circ denotes element-wise matrix multiplication operation.

(ii) Recovery by EBM in Data Space. With the initial recovered result Y^\hat{Y}^{\prime} generated by the coarse saliency predictor pαp_{\alpha}, the fine saliency predictor pθp_{\theta} can further refine the result by running a finite-step Langevin dynamics, which is initialized with Y^\hat{Y}^{\prime}, to obtain Y~\tilde{Y}^{\prime}. The underlying principle is that the initial recovery Y^\hat{Y}^{\prime} might be just around one local mode of the energy function. A few steps of Langevin dynamics (i.e., stochastic gradient descent) toward pθp_{\theta}, starting from Y^i\hat{Y}_{i}^{\prime}, will push Y^i\hat{Y}_{i}^{\prime} to its nearby low energy mode, which might correspond to its complete version YiY_{i}.

Algorithm 2 Cooperative learning while recovering

Input:
(1) Images {Xi}in\{X_{i}\}_{i}^{n} with incomplete annotations {Yi}in\{Y^{\prime}_{i}\}_{i}^{n};
(2) Number of learning iterations TT

Output: Parameters θ\theta and α\alpha

1:Initialize θ\theta and α\alpha with Gaussian noise
2:for t1t\leftarrow 1 to TT do
3:     Infer h^i\hat{h}^{\prime}_{i} from the visible part of YiY^{\prime}_{i} by Langevin dynamics in Eq. (7)
4:     Obtain initial recovery Y^i=Gα(Xi,h^i)\hat{Y}_{i}=G_{\alpha}(X_{i},\hat{h}^{\prime}_{i}).
5:     Revise Y^i\hat{Y}^{\prime}_{i} to obtain Y~i\tilde{Y}^{\prime}_{i} by Langevin dynamics in Eq. (2)
6:     Draw h^i𝒩(0,Id)\hat{h}_{i}\sim\mathcal{N}(0,I_{d})
7:     Sample initial prediction Y^i=Gα(Xi,h^i)\hat{Y}_{i}=G_{\alpha}(X_{i},\hat{h}_{i}).
8:     Revise Y^i\hat{Y}_{i} to obtain Y~i\tilde{Y}_{i} by Langevin dynamics in Eq. (2)
9:     Revise h^i\hat{h}_{i} to obtain h~i\tilde{h}_{i} by Langevin dynamics in Eq. (7)
10:     Update θ\theta with Δθ({Y~i},{Y~i})\Delta\theta(\{\tilde{Y}^{\prime}_{i}\},\{\tilde{Y}_{i}\}) using Adam
11:     Update α\alpha with Δα({h~i},{Y~i})\Delta\alpha(\{\tilde{h}_{i}\},\{\tilde{Y}_{i}\}) using Adam
12:end for

Cooperative Learning and Recovering. At each iteration tt, we perform the above cooperative recovery of the incomplete saliency maps {Y}i=1n\{Y^{\prime}\}_{i=1}^{n} via pθ(t)p_{\theta^{(t)}} and pα(t)p_{\alpha^{(t)}}, while learning pθ(t+1)p_{\theta^{(t+1)}} and pα(t+1)p_{\alpha^{(t+1)}} from {Xi,Y~i(t)}i=1n\{X_{i},\tilde{Y}_{i}^{\prime(t)}\}_{i=1}^{n}, where {Y~i(t)}i=1n\{\tilde{Y}_{i}^{\prime(t)}\}_{i=1}^{n} are the recovered saliency maps at iteration tt. The parameters θ\theta are still updated via Eq. (5) except that we replace YiY_{i} by Y~i\tilde{Y}^{\prime}_{i}. That is, at each iteration, we use the recovered Y~i\tilde{Y}^{\prime}_{i}, as well as the synthesized Y~i\tilde{Y}_{i}, to compute the gradient of the log-likelihood, which is denoted by Δθ({Y~i},{Y~i})\Delta\theta(\{\tilde{Y}^{\prime}_{i}\},\{\tilde{Y}_{i}\}). The algorithm simultaneously performs (i) cooperative recovering of missing annotations of each training example; (ii) cooperative sampling to generate annotations; (iii) cooperative learning of the two models by updating parameters with both recovered annotations and generated annotations. See Algorithm 2 for a detailed description of the learning while recovering algorithm.

5 Technical Details

We present the details of architecture designs of the LVM and the EBM, as well as the hyper-parameters below.

Latent Variable Model: The LVM Gα(X,h)G_{\alpha}(X,h), using the ResNet50 (He et al. 2016) as an encoder backbone, maps an image XX and a latent vector hh to a saliency map Y^\hat{Y}. Specifically, we adopt the decoder from the MiDaS (Ranftl et al. 2020) for its simplicity, which gradually aggregates the higher level features with lower level features via residual connections. We introduce the latent vector hh to the bottleneck of the LVM by concatenating the tiled hh with the highest level features of the encoder backbone, and then feed them to a 3×33\times 3 convolutional layer to obtain a feature map with the same size as the original highest level feature map of the encoder. The latent-vector-aware feature map is then fed to the decoder from Ranftl et al. (2020) to generate a final saliency map. As shown in Eq. (8), the parameters of the LVM are updated with the revised predictions {Y~}\{\tilde{Y}\} provided by the EBM. Thus, immature {Y~}\{\tilde{Y}\} in the early stage of the cooperative learning might bring in fluctuation in training the LVM, which in turn affects the convergence of the MCMC samples {Y~}\{\tilde{Y}\}. To stabilize the cooperative training, especially in the early stage, we let the LVM learn from not only {Y~}\{\tilde{Y}\} but also {Y}\{Y\}. Specifically, we add an extra loss for the LVM as λce(Gα(X,h~),Y)\lambda\mathcal{L}_{ce}(G_{\alpha}(X,\tilde{h}),Y), where λ\lambda linearly decreases to 0 during training, and ce\mathcal{L}_{ce} is the cross-entropy loss.

Energy-Based Model: The energy function Uθ(Y,X)U_{\theta}(Y,X) is parameterized by a neural network that maps the channel-wise concatenation of XX and YY to a scalar. Let cksl-n denote a k×\timesk Convolution-BatchNorm-ReLU layer with n filters and a stride of l. Let fc-n be a fully connected layer with n filters. The Uθ(Y,X)U_{\theta}(Y,X) is our framework consists of the following layers: c3s1-32, c4s2-64, c4s2-128, c4s2-256, c4s1-1, fc-100.

Table 1: Performance comparison with benchmark saliency prediction models, where “BkB” indicates the encoder backbone, and “R34” is ResNet34 backbone (He et al. 2016), and “R50” is the ResNet50 backbone (He et al. 2016).
DUTS ECSSD DUT HKU-IS PASCAL-S SOD
Method Year BkB SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow
Fully Supervised Models
PoolNet 2019 R50 .887 .840 .910 .037 .919 .913 .938 .038 .831 .748 .848 .054 .919 .903 .945 .030 .865 .835 .896 .065 .820 .804 .834 .084
BASNet 2019 R34 .876 .823 .896 .048 .910 .913 .938 .040 .836 .767 .865 .057 .909 .903 .943 .032 .838 .818 .879 .076 .798 .792 .827 .094
SCRN 2019 R50 .885 .833 .900 .040 .920 .910 .933 .041 .837 .749 .847 .056 .916 .894 .935 .034 .869 .833 .892 .063 .817 .790 .829 .087
F3Net 2020 R50 .888 .852 .920 .035 .919 .921 .943 .036 .839 .766 .864 .053 .917 .910 .952 .028 .861 .835 .898 .062 .824 .814 .850 .077
ITSD 2020 R50 .885 .840 .913 .041 .919 .917 .941 .037 .840 .768 .865 .061 .917 .904 .947 .031 .860 .830 .894 .066 .836 .829 .867 .076
LDF 2020 R50 .892 .861 .925 .034 .919 .923 .943 .036 .839 .770 .865 .052 .920 .913 .953 .028 .842 .768 .863 .064 - - - -
UCNet+ 2021 R50 .888 .860 .927 .034 .921 .926 .947 .035 .839 .773 .869 .051 .921 .919 .957 .026 .851 .825 .886 .069 .828 .827 .856 .076
PAKRN 2021 R50 .900 .876 .935 .033 .928 .930 .951 .032 .853 .796 .888 .050 .923 .919 .955 .028 .858 .838 .896 .067 .833 .836 .866 .074
Our_F 2021 R50 .902 .877 .936 .032 .928 .935 .955 .030 .857 .798 .889 .049 .927 .917 .960 .026 .873 .846 .909 .058 .854 .850 .885 .064
Weakly Supervised Models
SSAL 2020 R50 .803 .747 .865 .062 .863 .865 .908 .061 .785 .702 .835 .068 .865 .858 .923 .047 .798 .773 .854 .093 .750 .743 .801 .108
SCWS 2021 R50 .841 .818 .901 .049 .879 .894 .924 .051 .813 .751 .856 .060 .883 .892 .938 .038 .821 .815 .877 .078 .782 .791 .833 .090
Our_W 2021 R50 .847 .816 .902 .048 .896 .896 .934 .045 .817 .762 .864 .058 .894 .893 .943 .037 .834 .823 .886 .073 .803 .793 .849 .082
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Image GT SCRN F3Net ITSD LDF PAKRN Ours_F Uncertainty
Figure 1: Comparison of qualitative results of different fully supervised saliency prediction models.

Implementation Details: We train our model with a maximum of 30 epochs. Each image is rescaled to 352×352352\times 352. We set the number of dimensions of the latent space as d=8d=8. The number of Langevin steps is K=5K=5 and the Langevin step sizes for EBM and LVM are 0.4 and 0.1. The learning rates of the LVM and EBM are initialized to 5×1055\times 10^{-5} and 10310^{-3} respectively. We use Adam optimizer with momentum 0.9 and decrease the learning rates by 10% after every 20 epochs. It takes 20 hours to train the model with a batch size of 7 using a single NVIDIA GeForce RTX 2080Ti GPU.

6 Experiments

We conduct a series of experiments to test the performances of the proposed generative cooperative frameworks for saliency prediction. We start from experiment setup.

Datasets: We use the DUTS dataset (Wang et al. 2017) to train the fully supervised model, and S-DUTS (Zhang et al. 2020b) dataset with scribble annotations to train the weakly supervised model. Testing images include (1) DUTS testing dataset, (2) ECSSD (Yan et al. 2013), (3) DUT (Yang et al. 2013), (4) HKU-IS (Li and Yu 2015), (5) PASCAL-S (Li et al. 2014) and (6)SOD dataset (Movahedi and Elder 2010).

Compared methods: We compare our method against state-of-the-art fully supervised saliency detection methods, e.g., PoolNet (Liu et al. 2019), BASNet (Qin et al. 2019), SCRN (Wu, Su, and Huang 2019b), F3Net (Wei, Wang, and Huang 2020), ITSD (Zhou et al. 2020), LDF (Wei et al. 2020), UCNet+ (Zhang et al. 2021) and PAKRN (Xu et al. 2021). UCNet+ (Zhang et al. 2021) is the only generative framework. We also compare our weakly supervised solution with the scribble saliency detection models, e.g., SSAL (Zhang et al. 2020b) and SCWS (Yu et al. 2021).

Evaluation Metrics: We evaluate performance of our models and compared methods with four saliency evaluation metrics, including Mean Absolute Error (\mathcal{M}), mean F-measure (FβF_{\beta}), mean E-measure (EξE_{\xi}) (Fan et al. 2018) and S-measure (SαS_{\alpha}) (Fan et al. 2017).

Table 2: Experimental results of using EBMs as refinement modules and ablation study.
DUTS ECSSD DUT HKU-IS PASCAL-S SOD
SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow
EBM as Refinement Module
BASN_R .891 .842 .889 .041 .926 .921 .947 .035 .839 .781 .870 .051 .919 .925 .942 .031 .837 .749 .857 .070 .852 .859 .869 .090
SCRN_R .899 .857 .923 .034 .920 .921 .938 .037 .831 .748 .854 .053 .921 .919 .958 .027 .857 .769 .871 .062 .857 .867 .873 .080
Ablation Study
Gα(X)G_{\alpha}(X) .878 .835 .918 .038 .916 .915 .946 .036 .826 .751 .862 .058 .912 .901 .952 .030 .856 .830 .899 .064 .829 .827 .871 .072
Gα(X,h)G_{\alpha}(X,h) .897 .858 .932 .034 .918 .923 .946 .034 .837 .777 .882 .051 .914 .913 .957 .028 .863 .835 .900 .062 .831 .830 .874 .070
ITSD .885 .840 .913 .041 .919 .917 .941 .037 .840 .768 .865 .061 .917 .904 .947 .031 .860 .830 .894 .066 .836 .829 .867 .076
ITSD_Ours .914 .880 .945 .030 .938 .935 .959 .029 .860 .803 .901 .044 .933 .927 .971 .026 .875 .848 .921 .055 .845 .835 .880 .067
VGG16_Ours .906 .876 .941 .032 .939 .933 .953 .030 .857 .799 .893 .048 .929 .923 .959 .027 .871 .844 .907 .058 .841 .838 .871 .066
Our_F .902 .877 .936 .032 .928 .935 .955 .030 .857 .798 .889 .049 .927 .917 .960 .026 .873 .846 .909 .058 .854 .850 .885 .064
Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption
Image GT Scribble Recovered
Figure 2: Learning from images with scribble annotations. Each row shows one example of annotation recovery during training. The columns from left to right present training input images (“images”), ground truth annotations (“GT”) that are unknown for the learning algorithm, scribble annotations (“scribble”) as weak labels for training, and the output recovered annotations (“Recovered”) using the proposed “cooperative learning while recovering” strategy.

6.1 Fully Supervised Saliency Prediction

We first test the performance of our fully supervised generative cooperative saliency prediction framework.

Quantitative comparison: We compare the performance of our models and the compared methods in Table 1, where “Ours_F” denotes the proposed fully supervised models. We observe consistent performance improvement of “Ours_F” over six testing datasets compared with benchmark models, which clearly shows the advantage of our model. Note that, we adopt an existing decoder structure, i.e., MiDaS decoder (Ranftl et al. 2020), for the latent variable model in our proposed framework due to its easy implementation. We conduct an ablation study in Section 6.4 to further investigate the design of the decoder. Since our model uses a stochastic method, i.e., cooperative sampling, for prediction, we report the mean prediction to evaluate the performance of our models. Also, we observe relatively stable performance for different samples of predictions in larger testing datasets, e.g., DUTS testing dataset (Wang et al. 2017), and slightly fluctuant performance in smaller testing datasets, e.g., SOD (Movahedi and Elder 2010) testing dataset.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Image GT SSAL SCWS Ours_W Uncer.
Figure 3: Comparison of qualitative results obtained by different weakly supervised models learned from scribble annotations. Each row of example illustrates an input testing image, the corresponding ground truth saliency map as reference, the predictions from compared methods SSAL and SCWS, the mean prediction (“Ours_W”) and the uncertainty map (“Uncer.”) of our method .

Qualitative comparison: Figure 1 displays some qualitative results of the saliency predictions produced by our method and the compared methods. Each row corresponds to one example, and shows an input testing image, the corresponding ground truth saliency map, saliency maps predicted by SCRN, F3Net, ITSD, LDF and PAKRN, followed by the mean predicted saliency map and the pixel-wise uncertainty map of our model. The uncertainty map indicates the model confidence in predicting saliency from a given image, and is computed as the entropy (Kendall et al. 2017; Kendall and Gal 2017; Zhang et al. 2021, 2020a) of our model predictions. Results show that our method can not only produce visually reasonable saliency maps for input images but also meaningful uncertainty maps that are consistent with human perception.

Prediction time and model size comparison: We have two main modules in our framework, namely a latent variable model and an energy-based model. The former takes the ResNet50 (He et al. 2016) backbone as encoder, and the MiDaS (Ranftl et al. 2020) decoder for feature aggregation, leading to a model parameter size of 55M for the LVM. The latter adds 1M extra parameters to the cooperative learning framework. Thus, our model size is a total of 56M, which is comparable with mainstream saliency detection models, e.g., F3Net (Wei, Wang, and Huang 2020) has 48M parameters. As to the cooperative prediction time, it costs approximately 0.08 seconds to output a single prediction of saliency map, which is comparable with existing solutions as well.

6.2 Weakly Supervised Saliency Prediction

We then evaluate our weakly supervised generative cooperative saliency prediction framework on a dataset with scribble annotations (Zhang et al. 2020b), and show prediction performance on six testing sets in Table 1, where “Our_W” in the “Weakly Supervised Models” panel denotes our model. Figure 2 shows some examples of annotation recovery during training, where each row of example displays an input training image, the ground truth annotation as reference, scribble annotation used for training (the yellow scribble indicates the salient region, and the blue scribble indicates the background region), and the recovered saliency annotation obtained by our method. We compare our model with baseline methods, e.g., SSAL (Zhang et al. 2020b) and SCWS (Yu et al. 2021). The better performance of our model in testing shows the effectiveness of the proposed “cooperative learning while recovering” algorithm. Figure 3 displays a comparison of qualitative results obtained by different weakly supervised saliency prediction methods in testing.

6.3 Energy Function as a Refinement Module

As shown in Eq. (2), the EBM can iteratively refine the saliency prediction by Langevin sampling. With a well-trained energy function, we can treat it as a refinement module to refine predictions from existing saliency detection models. To demonstrate this idea, we select “BASN” (Qin et al. 2019) and “SCRN” (Wu, Su, and Huang 2019b) as base models due to the accessibility of their codes and predictions. We refine their predictions with the trained EBM and denote them by “BASN_R” and “SCRN_R”, respectively. Performances are shown in Table 2. Comparing with the performance of the base models in Table 1, we observe consistent performance improvements by using the trained EBM for refinement in Table 2. We show three examples of these models with and without EBM refinement in Figure 4. The qualitative improvement due to the usage of the EBM refinement verifies the usefulness of learned energy function.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Image GT BASN BASN_R SCRN SCRN_R
Figure 4: Comparison of qualitative results of base models without and with the trained EBM as a refinement module. “BASN” and “SCRN” are base models. “BASN_R” and “SCRN_R” are the corresponding methods with EBM refinement.

6.4 Ablation Study

We conduct the following experiments as shown in Table 2 to further analyze our proposed framework. Training a deterministic noise-free encoder-decoder GαG_{\alpha}: We remove the latent vector hh from our noise-injected encoder-decoder Gα(X,h)G_{\alpha}(X,h) and obtain a deterministic noise-free encoder-decoder Gα(X)G_{\alpha}(X). We train Gα(X)G_{\alpha}(X) with the cross-entropy loss as in those conventional deterministic saliency prediction models. In comparison with the state-of-the-art deterministic saliency detection model, i.e., PAKRN (Xu et al. 2021), Gα(X)G_{\alpha}(X) shows inferior performance due to its usage of a relatively small decoder (Ranftl et al. 2020). However, the superior performance of “Ours_F”, which is built upon Gα(X,h)G_{\alpha}(X,h) that shares the same decoder structure with Gα(X)G_{\alpha}(X), has exhibited the usefulness of the latent vector hh for generative modeling and verified the effectiveness of the EBM for cooperative learning.

Training a latent variable model Gα(X,h)G_{\alpha}(X,h) without the EBM: To further validate the importance of the cooperative training, we train a single latent variable model Gα(X,h)G_{\alpha}(X,h) without relying on an EBM, which leads to the alternating back-propagation training scheme (Han et al. 2017). Gα(X,h)G_{\alpha}(X,h) directly learns from observed training data rather than synthesized examples provided by an EBM. Compared with Gα(X,h)G_{\alpha}(X,h) trained independently, ”Ours_F“ achieves better performance, which validates the effectiveness of the cooperative learning strategy.

Table 3: Performance comparison with alternative uncertainty estimation methods.
DUTS ECSSD DUT HKU-IS PASCAL-S SOD
Method SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow SαS_{\alpha}\uparrow FβF_{\beta}\uparrow EξE_{\xi}\uparrow \mathcal{M}\downarrow
CVAE .890 .849 .925 .036 .919 .918 .948 .034 .836 .761 .868 .056 .918 .906 .955 .028 .863 .835 .902 .062 .838 .830 .878 .071
CGAN .888 .849 .927 .035 .917 .914 .944 .036 .837 .764 .871 .054 .917 .908 .955 .028 .865 .839 .906 .059 .836 .833 .874 .072
MCD .881 .842 .918 .038 .917 .917 .944 .036 .828 .753 .859 .057 .915 .908 .951 .030 .863 .837 .902 .062 .834 .831 .868 .073
ENS .885 .841 .921 .037 .921 .917 .948 .035 .831 .752 .862 .057 .916 .901 .952 .030 .858 .827 .897 .065 .835 .828 .872 .073
Our_F .902 .877 .936 .032 .928 .935 .955 .030 .857 .798 .889 .049 .927 .917 .960 .026 .873 .846 .909 .058 .854 .850 .885 .064
Refer to caption
Figure 5: Saliency predictions of alternative uncertainty estimation methods. For each panel, the first row shows an input image followed by the mean predictions of different alternative uncertainty estimation methods and ours, and the second row shows the ground truth saliency map followed by the uncertainty maps of different methods. From left to right columns, they are Image/ground truth, “CVAE”, “CGAN”, “MCD”, “ENS” and ours.

Design of encoder and encoder structures: We replace the decoder part in our proposed framework by the one of those existing deterministic saliency prediction methods, e.g., ITSD (Zhou et al. 2020). We select ITSD (Zhou et al. 2020) because of the availability of its code and the state-of-the-art performance. We show its performance in Table 2 as “ITSD_Ours”. Further, we replace the ResNet50 (He et al. 2016) encoder backbone in our model by VGG16 (Simonyan and Zisserman 2014) and denote the new model as “VGG16_Ours”. The consistently better performance of “ITSD_Ours” than the original “ITSD” validates the superiority of the generative cooperative learning framework. Comparable performances are observed in our models with different backbone selections.

6.5 Alternative Uncertainty Estimation Methods

In this section, we compare our generative framework with other alternative uncertainty estimation methods for saliency prediction. We first design two alternatives based on CVAEs (Sohn, Lee, and Yan 2015b) and CGANs (Mirza and Osindero 2014), respectively. For the CVAE model, we follow Zhang et al. (2021), except that we replace its decoder network by our decoder (Ranftl et al. 2020). As to the CGAN, we optimize the adversarial loss (Goodfellow et al. 2014) of the conditional generative adversarial network that consists of a conditional generator GG and a conditional discriminator DD. Specifically, we use the same latent variable model Gα(X,h)G_{\alpha}(X,h) as that in our model for the generator of CGAN. For the discriminator, we design a fully convolutional discriminator as in Hung et al. (2018) to classify each pixel into real (ground truth) or fake (prediction). To train CGAN, the discriminator DD is updated with the discriminator loss ce(D(Y),𝟏)+ce(D(Gα(X,h)),𝟎)\mathcal{L}_{ce}(D(Y),\mathbf{1})+\mathcal{L}_{ce}(D(G_{\alpha}(X,h)),\mathbf{0}), where ce\mathcal{L}_{ce} is the binary cross-entropy loss, 𝟏\mathbf{1} and 𝟎\mathbf{0} are all-one and all-zero maps of the same spatial size as YY. GαG_{\alpha} is updated with ce(Gα(X,h),Y)+λdce(D(Gα(X,h)),𝟏)\mathcal{L}_{ce}(G_{\alpha}(X,h),Y)+\lambda_{d}\mathcal{L}_{ce}(D(G_{\alpha}(X,h)),\mathbf{1}), where ce(D(Gα(X,h)),𝟏)\mathcal{L}_{ce}(D(G_{\alpha}(X,h)),\mathbf{1}) is the adversarial loss for GG and ce(Gα(X,h),Y)\mathcal{L}_{ce}(G_{\alpha}(X,h),Y) is the cross-entropy loss between the outputs of GG and the observed saliency maps. We set λd=0.1\lambda_{d}=0.1.

We also design two ensemble-based saliency detection frameworks with Monte Carlo dropout (Gal and Ghahramani 2016) and deep ensemble (Lakshminarayanan, Pritzel, and Blundell 2017) to produce multiple predictions, and show their performance as “MCD” and “ENS” in Table 3 respectively. For “MCD”, we add dropout to each level of features of the encoder within the noise-free encoder-decoder Gα(X)G_{\alpha}(X) with a dropout rate 0.3, and use dropout in both of training and testing processes. For “ENS”, we attach five MiDaS decoder (Ranftl et al. 2020) to Gα(X)G_{\alpha}(X), which are initialized differently, leading to five outputs of predictions. For both ensemble-based frameworks, similar to our generative models, we use the mean prediction averaging over 10 samples in testing as the final prediction, and the entropy of the mean prediction as the predictive uncertainty following Skafte, Jø rgensen, and Hauberg (2019).

We show performance of alternative uncertainty estimation models in Table 3, and visualize the mean prediction and predictive uncertainty for each method in Figure 5. For the CVAE-based framework, designing the approximate inference network takes extra efforts, and the imbalanced inference model may lead to the posterior collapse issue as discussed in He et al. (2019). For the CGAN-based model, according to our experiments, the training is sensitive to the proportion of the adversarial loss. Further, it cannot infer the latent variables hh, which makes the model hard to learn from incomplete data for weakly supervised learning. For the deep ensemble (Lakshminarayanan, Pritzel, and Blundell 2017) and MC dropout (Gal and Ghahramani 2016) solutions, they can hardly improve model performance, although the produced predictive uncertainty maps can explain model prediction to some extent. Compared all above alternative methods, our proposed framework is stable due to maximum likelihood learning, and we can infer latent variables hh without the need of an extra encoder. Further, as we directly sample from the truth posterior distribution via Langevin dynamics, instead of the approximated inference network, we have more reliable and accurate predictive uncertainty maps compared with other alternative solutions.

7 Conclusion and Discussion

In this paper, we propose a novel energy-based generative saliency prediction framework based on the conditional generative cooperative network, where a conditional latent variable model and an conditional energy-based model are jointly trained in a cooperative learning scheme to achieve a coarse-to-fine saliency prediction. The latent variable model serves as a coarse saliency predictor that provides a fast initial saliency prediction, while the energy-based model serves as a fine saliency predictor that further refines the initial output by the Langevin revision. Even though each of the models can represent the conditional probability distribution of saliency, the cooperative representation and training can offer the best of both worlds. Moreover, we propose a cooperative learning while recovering strategy and apply the model to the weakly supervised saliency detection scenario, in which partial annotations (e.g., scribble annotations) are provided for training. As to the cooperative recovery part of the proposed strategy, the latent variable model serves as a fast but coarse saliency recoverer that provides an initial recovery of the missing annotations from the latent space via inference process, while the energy-based model serves as a slow but fine saliency recoverer that refines the initial recovery results by Langevin dynamics. Combining these two types of recovery schemes leads to a coarse-to-fine recoverer. Further, we find that the learned energy function in the energy-based model can serve as a refinement module, which can be easily plugged into the existing pre-trained saliency prediction models. The energy function is the potential cost function trained from the saliency prediction task. In comparison to a mapping function from image to saliency, the cost function captures the criterion to measure the quality of the saliency given an image, and is more generalizable so that it can be used to refine other saliency predictions. Extensive results exhibit that, compared with both conventional deterministic mapping methods and alternative uncertainty estimation methods, our framework can lead to both accurate saliency predictions for computer vision tasks and reliable uncertainty maps indicating the model confidence in performing saliency prediction from an image. As to a broader impact, the proposed computational framework might also benefit the researchers in the field of computational neuroscience who investigate human attentional mechanisms. The proposed coarse-to-fine saliency prediction model and recovery model may shed light on a clear path toward the understanding of relationship between visual signals and human saliency.

References

  • Du and Mordatch (2019) Du, Y.; and Mordatch, I. 2019. Implicit Generation and Modeling with Energy Based Models. In NeurIPS, 3608–3618.
  • Fan et al. (2017) Fan, D.-P.; Cheng, M.-M.; Liu, Y.; Li, T.; and Borji, A. 2017. Structure-measure: A New Way to Evaluate Foreground Maps. In ICCV, 4548–4557.
  • Fan et al. (2018) Fan, D.-P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.-M.; and Borji, A. 2018. Enhanced-alignment Measure for Binary Foreground Map Evaluation. In IJCAI, 698–704.
  • Gal and Ghahramani (2016) Gal, Y.; and Ghahramani, Z. 2016. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In ICML, 1050–1059.
  • Gao et al. (2018) Gao, R.; Lu, Y.; Zhou, J.; Zhu, S.; and Wu, Y. 2018. Learning Generative ConvNets via Multi-grid Modeling and Sampling. In CVPR, 9155–9164.
  • Goodfellow et al. (2014) Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2014. Generative Adversarial Nets. In NeurIPS, 2672–2680.
  • Grathwohl et al. (2020) Grathwohl, W.; Wang, K.-C.; Jacobsen, J.-H.; Duvenaud, D.; Norouzi, M.; and Swersky, K. 2020. Your classifier is secretly an energy based model and you should treat it like one. In ICLR.
  • Han et al. (2017) Han, T.; Lu, Y.; Zhu, S.; and Wu, Y. 2017. Alternating Back-Propagation for Generator Network. In AAAI.
  • He et al. (2019) He, J.; Spokoyny, D.; Neubig, G.; and Berg-Kirkpatrick, T. 2019. Lagging Inference Networks and Posterior Collapse in Variational Autoencoders. In ICLR.
  • He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep Residual Learning for Image Recognition. In CVPR, 770–778.
  • Hung et al. (2018) Hung, W.-C.; Tsai, Y.-H.; Liou, Y.-T.; Lin, Y.-Y.; and Yang, M.-H. 2018. Adversarial Learning for Semi-supervised Semantic Segmentation. In BMVC.
  • Itti, Koch, and Niebur (1998) Itti, L.; Koch, C.; and Niebur, E. 1998. A Model of Saliency-based Visual Attention for Rapid Scene Analysis. TPAMI, 20: 1254 – 1259.
  • Jimenez Rezende, Mohamed, and Wierstra (2014) Jimenez Rezende, D.; Mohamed, S.; and Wierstra, D. 2014. Stochastic Backpropagation and Approximate Inference in Deep Generative Models. In ICML.
  • Kendall et al. (2017) Kendall, A.; Badrinarayanan, V.; ; and Cipolla, R. 2017. Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding. In BMVC.
  • Kendall and Gal (2017) Kendall, A.; and Gal, Y. 2017. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? In NeurIPS.
  • Kingma and Welling (2014) Kingma, D.; and Welling, M. 2014. Auto-Encoding Variational Bayes. In ICLR.
  • Kingma and Ba (2015) Kingma, D. P.; and Ba, J. 2015. Adam: A Method for Stochastic Optimization. In International Conference on Learning Representations (ICLR).
  • Kingma and Welling (2013) Kingma, D. P.; and Welling, M. 2013. Auto-Encoding Variational Bayes. In ICLR.
  • Kohl et al. (2018) Kohl, S. A.; Romera-Paredes, B.; Meyer, C.; De Fauw, J.; Ledsam, J. R.; Maier-Hein, K. H.; Eslami, S.; Rezende, D. J.; and Ronneberger, O. 2018. A Probabilistic U-Net for Segmentation of Ambiguous Images. NeurIPS.
  • Lakshminarayanan, Pritzel, and Blundell (2017) Lakshminarayanan, B.; Pritzel, A.; and Blundell, C. 2017. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles. In NeurIPS. Curran Associates, Inc.
  • Li, Xie, and Lin (2018a) Li, G.; Xie, Y.; and Lin, L. 2018a. Weakly Supervised Salient Object Detection Using Image Labels. In AAAI.
  • Li, Xie, and Lin (2018b) Li, G.; Xie, Y.; and Lin, L. 2018b. Weakly Supervised Salient Object Detection Using Image Labels. In AAAI.
  • Li and Yu (2015) Li, G.; and Yu, Y. 2015. Visual saliency based on multiscale deep features. In CVPR, 5455–5463.
  • Li et al. (2014) Li, Y.; Hou, X.; Koch, C.; Rehg, J. M.; and Yuille, A. L. 2014. The Secrets of Salient Object Segmentation. In CVPR, 280–287.
  • Liu et al. (2019) Liu, J.-J.; Hou, Q.; Cheng, M.-M.; Feng, J.; and Jiang, J. 2019. A Simple Pooling-Based Design for Real-Time Salient Object Detection. In CVPR.
  • Liu, Han, and Yang (2018) Liu, N.; Han, J.; and Yang, M.-H. 2018. PiCANet: Learning Pixel-wise Contextual Attention for Saliency Detection. In CVPR, 3089–3098.
  • Luc et al. (2016) Luc, P.; Couprie, C.; Chintala, S.; and Verbeek, J. 2016. Semantic Segmentation using Adversarial Networks. In NeurIPS Workshop on Adversarial Training.
  • Mirza and Osindero (2014) Mirza, M.; and Osindero, S. 2014. Conditional Generative Adversarial Nets. CoRR, abs/1411.1784.
  • Movahedi and Elder (2010) Movahedi, V.; and Elder, J. H. 2010. Design and perceptual validation of performance measures for salient object segmentation. In CVPR Workshop, 49–56.
  • Neal (2012) Neal, R. 2012. MCMC using Hamiltonian dynamics. Handbook of Markov Chain Monte Carlo.
  • Nguyen et al. (2019) Nguyen, D. T.; Dax, M.; Mummadi, C. K.; Ngo, T.-P.-N.; Nguyen, T. H. P.; Lou, Z.; and Brox, T. 2019. DeepUSPS: Deep Robust Unsupervised Saliency Prediction With Self-Supervision. In NeurIPS.
  • Nijkamp et al. (2019) Nijkamp, E.; Hill, M.; Zhu, S.-C.; and Wu, Y. N. 2019. Learning Non-Convergent Non-Persistent Short-Run MCMC Toward Energy-Based Model. NeurIPS.
  • Pan et al. (2017) Pan, J.; Canton, C.; McGuinness, K.; O’Connor, N. E.; Torres, J.; Sayrol, E.; and Giro-i Nieto, X. a. 2017. SalGAN: Visual Saliency Prediction with Generative Adversarial Networks. In CVPR Workshop.
  • Qin et al. (2019) Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; and Jagersand, M. 2019. BASNet: Boundary-Aware Salient Object Detection. In CVPR, 7479–7489.
  • Ranftl et al. (2020) Ranftl, R.; Lasinger, K.; Hafner, D.; Schindler, K.; and Koltun, V. 2020. Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer. TPAMI.
  • Simonyan and Zisserman (2014) Simonyan, K.; and Zisserman, A. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. In ICLR.
  • Skafte, Jø rgensen, and Hauberg (2019) Skafte, N.; Jø rgensen, M.; and Hauberg, S. r. 2019. Reliable training and estimation of variance networks. In NeurIPS.
  • Sohn, Lee, and Yan (2015a) Sohn, K.; Lee, H.; and Yan, X. 2015a. Learning Structured Output Representation using Deep Conditional Generative Models. In NeurIPS, 3483–3491.
  • Sohn, Lee, and Yan (2015b) Sohn, K.; Lee, H.; and Yan, X. 2015b. Learning Structured Output Representation using Deep Conditional Generative Models. In NeurIPS, 3483–3491.
  • Souly, Spampinato, and Shah (2017) Souly, N.; Spampinato, C.; and Shah, M. 2017. Semi Supervised Semantic Segmentation Using Generative Adversarial Network. In ICCV, 5689–5697.
  • Wang et al. (2017) Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; and Ruan, X. 2017. Learning to detect salient objects with image-level supervision. In CVPR, 136–145.
  • Wang et al. (2018) Wang, T.; Zhang, L.; Wang, S.; Lu, H.; Yang, G.; Ruan, X.; and Borji, A. 2018. Detect Globally, Refine Locally: A Novel Approach to Saliency Detection. In CVPR, 3127–3135.
  • Wang et al. (2019) Wang, W.; Shen, J.; Cheng, M.-M.; and Shao, L. 2019. An Iterative and Cooperative Top-Down and Bottom-Up Inference Network for Salient Object Detection. In CVPR.
  • Wang et al. (2019) Wang, W.; Zhao, S.; Shen, J.; Hoi, S. C. H.; and Borji, A. 2019. Salient Object Detection With Pyramid Attention and Salient Edges. In CVPR, 1448–1457.
  • Wei, Wang, and Huang (2020) Wei, J.; Wang, S.; and Huang, Q. 2020. F3Net: Fusion, Feedback and Focus for Salient Object Detection. In AAAI.
  • Wei et al. (2020) Wei, J.; Wang, S.; Wu, Z.; Su, C.; Huang, Q.; and Tian, Q. 2020. Label Decoupling Framework for Salient Object Detection. In CVPR, 13025–13034.
  • Wu, Su, and Huang (2019a) Wu, Z.; Su, L.; and Huang, Q. 2019a. Cascaded Partial Decoder for Fast and Accurate Salient Object Detection. In CVPR, 3907–3916.
  • Wu, Su, and Huang (2019b) Wu, Z.; Su, L.; and Huang, Q. 2019b. Stacked Cross Refinement Network for Edge-Aware Salient Object Detection. In ICCV.
  • Xie et al. (2018a) Xie, J.; Lu, Y.; Gao, R.; Zhu, S.-C.; and Wu, Y. N. 2018a. Cooperative Training of Descriptor and Generator Networks. TPAMI.
  • Xie et al. (2016) Xie, J.; Lu, Y.; Zhu, S.-C.; and Wu, Y. 2016. A Theory of Generative ConvNet. In ICML, volume 48, 2635–2644.
  • Xie et al. (2021a) Xie, J.; Xu, Y.; Zheng, Z.; Zhu, S.; and Wu, Y. N. 2021a. Generative PointNet: energy-based learning on unordered point sets for 3D generation, reconstruction and classification. In CVPR.
  • Xie et al. (2021b) Xie, J.; Zheng, Z.; Fang, X.; Zhu, S.-C.; and Wu, Y. N. 2021b. Cooperative Training of Fast Thinking Initializer and Slow Thinking Solver for Conditional Learning. TPAMI.
  • Xie et al. (2020) Xie, J.; Zheng, Z.; Gao, R.; Wang, W.; Zhu, S.; and Wu, Y. 2020. Generative VoxelNet: learning energy-based models for 3D shape synthesis and analysis. TPAMI.
  • Xie et al. (2018b) Xie, J.; Zheng, Z.; Gao, R.; Wang, W.; Zhu, S.-C.; and Nian Wu, Y. 2018b. Learning descriptor networks for 3D shape synthesis and analysis. In CVPR, 8629–8638.
  • Xie, Zheng, and Li (2021) Xie, J.; Zheng, Z.; and Li, P. 2021. Learning Energy-Based Model with Variational Auto-Encoder as Amortized Sampler. In AAAI.
  • Xie, Zhu, and Wu (2019) Xie, J.; Zhu, S.-C.; and Wu, Y. N. 2019. Learning energy-based spatial-temporal generative convnets for dynamic patterns. TPAMI.
  • Xu et al. (2021) Xu, B.; Liang, H.; Liang, R.; and Chen, P. 2021. Locate Globally, Segment Locally: A Progressive Architecture With Knowledge Review Network for Salient Object Detection. In AAAI, 3004–3012.
  • Xue et al. (2017) Xue, Y.; Xu, T.; Zhang, H.; Long, R.; and Huang, X. 2017. SegAN: Adversarial Network with Multi-scale L1L_{1} Loss for Medical Image Segmentation. Neuroinformatics, 16.
  • Yan et al. (2013) Yan, Q.; Xu, L.; Shi, J.; and Jia, J. 2013. Hierarchical saliency detection. In CVPR, 1155–1162.
  • Yang et al. (2013) Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; and Yang, M.-H. 2013. Saliency Detection via Graph-Based Manifold Ranking. In CVPR, 3166–3173.
  • Yu and Cai (2018) Yu, H.; and Cai, X. 2018. Saliency detection by conditional generative adversarial network. In Ninth International Conference on Graphic and Image Processing, 253.
  • Yu et al. (2021) Yu, S.; Zhang, B.; Xiao, J.; and Lim, E. G. 2021. Structure-Consistent Weakly Supervised Salient Object Detection with Local Saliency Coherence. In AAAI.
  • Zhang, Han, and Zhang (2017) Zhang, D.; Han, J.; and Zhang, Y. 2017. Supervision by Fusion: Towards Unsupervised Learning of Deep Salient Object Detector. In ICCV, 4068–4076.
  • Zhang et al. (2021) Zhang, J.; Fan, D.-P.; Dai, Y.; Anwar, S.; Saleh, F.; Aliakbarian, S.; and Barnes, N. 2021. Uncertainty Inspired RGB-D Saliency Detection. TPAMI.
  • Zhang et al. (2020a) Zhang, J.; Fan, D.-P.; Dai, Y.; Anwar, S.; Saleh, F. S.; Zhang, T.; and Barnes, N. 2020a. UC-Net: Uncertainty Inspired RGB-D Saliency Detection via Conditional Variational Autoencoders. In CVPR.
  • Zhang et al. (2020b) Zhang, J.; Yu, X.; Li, A.; Song, P.; Liu, B.; and Dai, Y. 2020b. Weakly-Supervised Salient Object Detection via Scribble Annotations. In CVPR.
  • Zhang et al. (2018) Zhang, J.; Zhang, T.; Dai, Y.; Harandi, M.; and Hartley, R. 2018. Deep Unsupervised Saliency Detection: A Multiple Noisy Labeling Perspective. In CVPR, 9029–9038.
  • Zhang et al. (2018) Zhang, X.; Zhu, X.; Zhang, . X.; Zhang, N.; Li, P.; and Wang, L. 2018. SegGAN: Semantic Segmentation with Generative Adversarial Network. In 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), 1–5.
  • Zhao, Xie, and Li (2021) Zhao, Y.; Xie, J.; and Li, P. 2021. Learning Energy-Based Generative Models via Coarse-to-Fine Expanding and Sampling. In International Conference on Learning Representations (ICLR).
  • Zheng, Xie, and Li (2021) Zheng, Z.; Xie, J.; and Li, P. 2021. Patchwise Generative ConvNet: Training Energy-Based Models From a Single Natural Image for Internal Learning. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2961–2970.
  • Zhou et al. (2020) Zhou, H.; Xie, X.; Lai, J.-H.; Chen, Z.; and Yang, L. 2020. Interactive Two-Stream Decoder for Accurate and Fast Saliency Detection. In CVPR.
  • Zhu et al. (2017) Zhu, J.-Y.; Zhang, R.; Pathak, D.; Darrell, T.; Efros, A. A.; Wang, O.; and Shechtman, E. 2017. Toward Multimodal Image-to-Image Translation. In NeurIPS, 465–476.