Renderers are Good Zero-Shot Representation Learners: Exploring Diffusion Latents for Metric Learning
Abstract
Can the latent spaces of modern generative neural rendering models serve as representations for 3D-aware discriminative visual understanding tasks? We use retrieval as a proxy for measuring the metric learning properties of the latent spaces of Shap-E [10], including capturing view-independence and enabling the aggregation of scene representations from the representations of individual image views, and find that Shap-E representations outperform those of the classical EfficientNet baseline representations zero-shot, and is still competitive when both methods are trained using a contrative loss. These findings give preliminary indication that 3D-based rendering and generative models can yield useful representations for discriminative tasks in our innately 3D-native world. Our code is available at https://github.com/michaelwilliamtang/golden-retriever.
1 Introduction
Recent advances in neural rendering have enabled 2D novel views and entire 3D models to be generated from as few as a single input 2D view, with incredibly high quality [13, 15, 17]. Early work on novel view synthesis, focusing on neural radiance fields, takes a prior-agnostic approach, training a feature volume on many views of a single scene at inference time before using an algorithmic rendering equation to sample density and color predictions along rays in the novel view to construct it [15]. More recently, both neural radiance field and diffusion-based methods take a prior-based approach, leveraging large datasets of 2D and 3D scenes to learn priors, and computing novel views or 3D models of a held-out scene by performing forward passes through the model at inference time [13, 25]. This confers the advantage of being able to infer on less input views, which removes cumbersome inference-time data collection constraints, as well as eliminates the need to undergo multi-hour training procedures at inference time (although recent work has found various optimizations that drastically reduce this training process [12, 16]). From an information theoretic perspective, learning 3D-aware priors on a large dataset captures a large amount of geometric information which can significantly reduce the difficulty of the novel view synthesis task, and more closely resembles the conceptual view synthesis processes of humans. For complex scenes, where it is geometrically impossible to capture all of the details from only a handful of camera views, these priors are incredibly important — this means downstream application users can focus on imaging the most important or notable parts of a scene while letting this latent knowledge fill in the more mundane details (e.g. the bottom of a chair or table). For these reasons, we focus on prior-based approaches.
Although the origin of these new methods is in generative tasks, a large number of vision and graphics applications involve discriminative understanding – for example, robotics applications rely on vision models as world models, a way of capturing camera data that leads to sophisticated, natural understanding of scenes, directly using those visual features as inputs into reinforcement learning and control modules [2], and visual-question-answering models may require strong 3D-aware priors to answer questions that discuss the geometry of a 2D input image, without any explicit generation outputs [1].
We posit that learning strong priors for generative tasks also leads to good learned 3D-aware representations. Specifically, we examine view independence as a proxy for 3D-awareness in the latent representations of a neural rendering model (Shap-E [10]) and a classical neural 2D image understanding model (EfficientNet [20]). We investigate this by constructing a 2D image retrieval task using different views of ShapeNet [3] car scenes, and evaluate the metric learning property of the embeddings produced by different models: a model induces a useful representation if the embeddings of different views of the same scene are mapped closer in vector space compared to views of different scenes.
Main contributions. Our main contributions are summarized below:
-
•
We constructed a dataset derived from ShapeNet [3] containing renders of 3D models, which is designed to explore 3-D awareness in metric learning via scene retrieval.
-
•
Performed zero-shot retrieval using embeddings from pre-trained image encoders, and showed that latent embeddings from 3D diffusion outperform those of classification CNNs.
-
•
Demonstrated that contrastive learning improves scene retrieval and learns an embedding with (imperfect) view invariance.
-
•
Report ablations showing dynamics of varying number of reference images per scene as well as number of images seen during training.
2 Related Work
2.1 Neural rendering for discriminative tasks
We note that the space of using generative (e.g. diffusion) approaches to solve discriminative tasks is remarkably underexplored. Mainly, a handful of works propose to use 2D diffusion models to solve specific dense prediction, learning to denoise intermediate artifacts like bounding boxes, depth maps, and filters by reversing a manually constructed noising process [5, 8, 9, 23, 24].
Notably, there have been no works analyzing 3D neural rendering models for discriminative tasks, and simultaneously no works directly analyzing using the strong visual priors of generative models using lens of representation learning. We aim to provide preliminary explorations on both of these areas, investigating a 3D diffusion model trained on a large dataset of 3D scenes, and analyzing one measurable aspect of the resultant 3D-aware priors — view independence. Our results encourage further exploration of this field and show that 3D diffusion models indeed induce useful representations zero-shot.
2.2 Metric learning for image retrieval
The state-of-the-art on 2D metric learning retrieval benchmarks is dominated by classic convolutional models trained on the metric learning objective using increasingly sophisticated contrastive loss functions. After the advent of triplet loss [18], which explicitly penalizes the distance between embeddings of same-labeled examples and rewards the distance (up to a threshold) between embeddings of different-labeled examples, various improvements were proposed that improve separability, including adding regularization terms relating to clustered classes tightly around class centers [21], normalizing the distances between such centers [14], and enforcing the decision margin in a more natural way (i.e. in angle space) [7]. Since we focus on preliminary investigations on the choice of the pre-trained backbone model rather than the choice of subsequent contrastive training, we use the simple SimCLR contrastive training setup with cross-entropy loss (which was determined in the paper to outperform triplet loss), with details in Section 2.5.
2.3 EfficientNet
EfficientNet [20] is a popular convolutional neural network baseline for a variety of scene understanding tasks, including achieving the state-of-the-art on transfer learning datasets such as CIFAR-100 [11] and ImageNet-1K [6]. It is also commonly used as a backbone for the metric learning models described in Section 2.2. Its main contribution was decreasing the number of parameters by an order of magnitude while producing significantly improved performance using observations about optimal scaling laws for CNNs (i.e., finding the relationship between scaling width, depth, and resolution). Here, we use the second-to-last layer of the network as the canonical latent representation for discriminative tasks.
2.4 Shap-E
Shap-E [10] is a state-of-the-art conditional generative model that generates 3D assets conditioned on 2D image or text by learning implicit function representations. Specifically, the work trains an encoder that maps 3D assets into latent codes representing the parameters of an implicit function, then trains a conditional diffusion model on these latent codes — we use the learned latent codes as the natural learned representation for image and scene understanding tasks.
2.5 SimCLR
To understand the effect of self-supervised training on the retrieval database (i.e., learning using the supervision that different views of the same database image belong to the same scene), we apply the SimCLR framework [4] for contrastive learning. SimCLR proposed a simple method for minimizing contrastive loss in the latent space with the following steps: [22]
-
1.
Sample a random data augmentation (e.g. random flip, crop, resize) to produce pairs of images that represent different views the same image, , for a given batch.
-
2.
Compute the embedding of each (possibly transformed) image by passing it through an encoder and a projection head.
-
3.
Treating as a positive pair and all other combinatorial pairs of (possibly transformed) images within the batch as in-batch negatives, minimize the contrastive loss over cosine similarity .
where is a temperature parameter.
The final representations are then the encoder functions (without the projection head), which can then be used to compute representation vectors for possibly held-out images on downstream tasks. In our setup, we treat different views of the same scene as the data augmentation in SimCLR, since we wish our embedding to cluster these views together.
3 Methods
3.1 Retrieval task
To understand whether the latent space of the Shap-E [10] neural rendering model better captures 3D-aware properties compared to that of classical convolutional models like EfficientNet [20], we can evaluate the view independence of each encoder via the metric learning property, where views of the same image should be mapped to vectors with high similarity. Formally, we can measure this for each encoder by computing its embedding for each input view of each scene (indicating this is view of scene ), where we would like it to be the case that
for all and all . (i.e., any pair of same-scene views should be closer than any pair of views from different scenes) Here, is a similarity function such as cosine similarity.
We can measure this by constructing the canonical 2D image retrieval task, where we construct a database of different scenes, with different image views each, and an query set of of those scenes with different image views each. The retrieval task is as follows: given a query image, return the most relevant scene from the database. The formulation that evaluates metric learning properties imposes the following retrieval algorithm:
-
1.
Compute for the query image .
-
2.
Compute for each view of each database scene.
-
3.
Compute the similarities for all according to similarity function , and return the top retrieved image
If we want the top images, this corresponds to retrieved images
(note that some images can be of the same scene).
This is depicted graphically in Figure 1.
In particular, the second — most costly — step can be precomputed in a single indexing step instead of before each new query image, and the third step can be implemented via highly optimized Maximum Inner Product Search algorithms, which makes this method highly efficient. This is notably the state-of-the-art algorithm for retrieval pipelines for dense retrieval methods across many domains, e.g. for text document retrievals. Thus to evaluate the view independence of the representations of different models, we substitute them in as different retrievers using this view-based dataset construction, and evaluate the resulting retrieval performance using the following standard metrics averaged over queries:
-
1.
Mean Reciprocal Rank (MRR)@: among the top retrieved images, if none are from the ground truth scene, return 0. Else return , where is the smallest index of a candidate that is from the ground truth scene. In this paper, unless otherwise specified, we choose (i.e. we find the rank over all predictions) and simply denote this as MRR.
-
2.
Recall@: among the top retrieved images, return the fraction of them that are from the ground truth scene.
Importantly, the query and database image views for each scene are disjoint in task construction, and in the zero-shot setting neither set is available at training time, mirroring real-world applications where users can upload images tagged with the scene they are related to (e.g., the entity name of a global landmark), as well as input queries where they are interested in retrieving the scene for a given new image (e.g., a new picture they took of some unknown global landmark in the database). Rather, the encoder must learn image understanding (and in particular 3D-aware ones such as view independence) by being trained on other datasets. Note that unfortunately, the exact Shap-E dataset is unreleased, so it is possible the scenes from our retrieval dataset comprises a small part of the their training data during the generative training process, but we specifically refrain from applying any discriminative training on any of the retrieval encoders during zero-shot evaluation.
Implementation. We used the validation set of the cars single-category subset of the ShapeNet dataset [3] ( scenes), and unless otherwise stated used , mimicking the real-life lack of more views available in the database but typically only one per query. We extracted latent spaces from both models as detailed in Section 2, using a PyTorch reimplementation of SimCLR [19]. We use cosine similarity in all experiments.

3.2 View aggregation
Some retrieval settings enforce the use of a single vector representation per scene (e.g. at large scales, for computational efficiency). Thus we also examine whether the latent representations of neural-rendering-based and classical models have other semantic properties, such as the ability to be aggregated under a mean. To evaluate this, we run a modified retrieval pipeline where the vector representation of a scene is the elementwise mean of the vector representations of each image view of that scene available in the database:
and then compare the resulting scene representations using the similiarity function to get top scenes. We denote this as the mean-aggregated setting.
3.3 Training on the database images
Alongside approaching the task zero-shot, we are also interested in how well the latent spaces adapt to the evaluation database, i.e., if models are given training-time access to the database images. Notably, there is still no training-time access to the query images. This emulates the real-world scenario where a large database of scenes of interest is uploaded (e.g., famous world landmarks), and an indexing service can perform further training on the encoder using the fixed database before deployment, although the potential query images users are interested in are arbitrary.
To do so, we trained the latent embeddings of Shap-E contrastively using the SimCLR [4] setup (details in Section 2) using randomly sampled views of the same scene as positive pairs and using in-batch negatives.
Implementation. We froze the encoders and trained a multi-layer (3 layers for EfficientNet and 5 for Shap-E performed best) fully-connected MLP as well as 1-layer projection head MLP on top using the contrastive objective, Following the original SimCLR paper, we discarded the 1-layer projection head during evaluation and treated the layer beneath as the representation. The depth and width were chosen using a hyperparameter sweep.
4 Results
4.1 Image Retrieval
Recall@1 | Recall@5 | MRR | |
---|---|---|---|
w/o learned encoder | |||
EfficientNet | 0.82 | 0.86 | 0.85 |
Shap-E | 0.86 | 0.96 | 0.90 |
w/ learned encoder | |||
EfficientNet | 0.94 | 1.00 | 0.96 |
Shap-E | 0.88 | 0.94 | 0.91 |
Recall@1 | Recall@5 | MRR | |
---|---|---|---|
w/o learned encoder | |||
EfficientNet | 0.62 | 0.76 | 0.70 |
Shap-E | 0.42 | 0.74 | 0.56 |
w/ learned encoder | |||
EfficientNet | 0.88 | 0.98 | 0.92 |
Shap-E | 0.82 | 0.98 | 0.88 |
We report metric learning retrieval results in Table 2, including the performance of both the zero-shot raw and contrastively trained EfficientNet and Shap-E embeddings. We visualize the effect of training by showing t-SNE plots of EfficientNet and Shap-E embeddings before and after training in Figure 4 as well as report just the density of those t-SNE plots to more cleanly visualize the separability of the clusters of images corresponding to different scenes in Figure 5. For these results, we used and trained on the same 20 images.
Overall, we find that Shap-E embeddings substantially outperform the EfficientNet embeddings zero-shot on all retrieval metrics, although SimCLR training improves both representations, including making EfficientNet embeddings competitive with the Shap-E embeddings. This is corroborated by the t-SNE visualizations, which show the Shap-E embeddings to be slightly more separated, and the trained versions of both embeddings to be much more separated, with the query embeddings almost always being mapped on top of a cluster of database embeddings (visualized as each star symbol being mapped on top of the circle symbols of the same color). The density t-SNE plots also support this, with the trained embeddings clearly better separated and sparser with higher peaks than their untrained counterparts.
We conjecture that the Shap-E latent space is particularly well-suited to cosine similarity because it was trained to reconstruct 3D meshes from input images; similar images should be encoded in similar feature vectors. EfficientNet, however, is trained to classify images in the diverse ImageNet [6] dataset. The feature space learned by EfficientNet’s penultimate layer must be separable by the final fully connected layer, but is not necessarily endowed with a similarity metric. However, this metric property is explicitly optimized by SimCLR, which is likely what allows both sets of learned embeddings to be strong.
4.2 Mean Aggregation
We report metric learning retrieval results in the mean-aggregated setting in Table 3. For this result, we used the same properties as in our main results (), but take represent each scene by taking the mean over representative vectors. Overall, we find that mean-aggregation degrades the performance of zero-shot retrieval on both the Shap-E and Efficient embeddings. However, contrastive learning dramatically improves the performance of mean-aggregation, and the mean-aggregated retrieval results after training are competitive with ordinary retrieval on the trained embeddings in Table 2. This suggests that by pre-processing our database with contrastive learning and per-scene mean-aggregation, we can speed up similarity search on future queries without sacrificing recall.
5 Analysis
We examine the effect of changing the number of views per scene available in the database at query time, , as well as the number of views per scene available as positive pairs in SimCLR training, which we will denote .
5.1 Effect of on retrieval
We plot evaluation results on databases with different values of for both trained and untrained (using a dotted line) EfficientNet embeddings, in both the standard and mean-aggregated retrieval settings in Figure 6, and the same for Shap-E embeddings in Figure 7. For these experiments, we use a fixed images per scene in training.
Note that in order to evaluate the robustness of the SimCLR-trained embeddings, alongside evaluating with a database of images including some seen at training time, we also consider the alternative setting where we evaluate on a disjoint database of unseen images. In the latter case, the training images, database images, and query image are all mutually disjoint for every scene. In this scenario, we only see good results if the training process learns general view-independent properties of the scene, instead of memorizing the specific positive pair images. Since we in fact find that the unseen and seen performances are comparable, we conclude the SimCLR contrastive learning process learns to encode view independence robustly.
Standard (un-aggregated) setting. We find that more views improves zero-shot embedding performance for both kinds of representations, with major improvements up to about 4 views for EfficientNet and higher (about 8) for Shap-E. One interpretation of this is that view independence has not fully been captured by either encoder, as seeing more views of a scene adds a large amount of information compared to what would be expected if the model fully extrapolated from views seen and only learned information from new views about previously occluded, asymmetric parts of the car (since the cars are highly symmetric this is expected to be small).
In contrast, training mostly removes this dependence when encoding both seen and unseen database images, which points to SimCLR being an effective training method for instilling view independence.
Mean-aggregated setting. Zero-shot, we find that both encoders admit some ability to represent scenes using mean aggregation on representations of individual views, as performance improves with more views being aggregated, pointing to new information being captured to some degree. Unfortunately, the raw performance under multiple retrieval metrics for both EfficientNet and Shap-E is poor.
However, we find that training consistently improves performance for both encoders in this setting. In accordance with our previous findings in the standard setting, the trained encoders have little dependence on in the aggregate setting. This reinforces the robustness of training, but also points to the fact that we may not be able to draw strong conclusions about whether the trained embeddings admit aggregation, since they already encode some degree of view independence and do not significantly benefit from more views — for example, in the ideal case where training endows full view independence (which we do not achieve here), aggregation would be between identical embeddings and thus be trivial, so some version of this effect is expected. It suffices to note that since both encoders receive a boost in base performance on a single view when trained and aggregating it with more images does not hurt performance for EfficientNet and slightly improves it for Shap-E, the training process slightly promotes the ability for representations to be aggregated. This is itself a positive result, since aggregation leads to significant computational and efficiency gains for downstream retrieval applications, so we conclude that the training process can significantly improve the EfficientNet and Shap-E representations (via improving base performance) while also preserving those practical gains (when aggregating additional image representations).
5.2 Effect of on retrieval
We evaluate metric learning retrieval performance for trained EfficientNet and Shap-E embeddings using different values of . We again consider unseen and seen settings for the trained encoders, and include the zero-shot, untrained performance of both encoders as a baseline, shown as a dotted line. For these experiments, we fix .
We see that training with more views indeed intuitively leads to substantially better performance for EfficientNet embeddings. In contrast, there is no clear pattern for Shap-E. We see some instability in training runs for intermediate values of , and for very low values of , the trained Shap-E actually often does worse at the metric learning retrieval task than its zero-shot counterpart. This may imply that it is hard to learn view independence in a many-layer MLP head on top of a high-dimensional Shap-E model with only a few views per scene, and in particular that the optimization landscape for training on top of Shap-E is less amenable to learning view independence than the one for training on top of EfficientNet, despite the fact that the zero-shot performance of Shap-E is higher.
Again, we find that unseen and seen performance across the board is about the same, pointing to SimCLR training being robust.
6 Discussion
The main limitation of this work is that it focuses on a single domain (synthetic Blender cars from ShapeNet [3]) and one neural-rendering-based encoders, namely Shap-E [10]. We find that this encoder exhibits useful properties as a representation learner, opening up a new area for research, but the result would be stronger if it were replicated across multiple different encoders, especially given the diversity of different neural rendering approaches, including non-diffusion-based ones. Replicating this across multiple large benchmarks and a high-quality in-the-wild dataset would also significantly strengthen the results Under ideal circumstances, we would also want access to the details of the training set of Shap-E to make sure it is definitively disjoint from our retrieval task.
Another limitation of this work is the large memory footprint of the proposed vector embeddings. Each Shap-E embedding is a -dimensional vector, requiring MB per embedding or MB per scene (with ) to store in memory. For large retrieval databases, this memory consumption may be too costly to achieve zero-shot Shap-E retrieval at scale. After training, however, each learned embedding takes on the order of kilobytes to store, suggesting that such a database may scale after pre-processing with SimCLR and mean-aggregation.
7 Conclusion
We conclude that neural-rendering-based generative models induce useful representations in latent space, encoding images in a way that captures view independence as well as the ability to aggregate views to form a single coherent scene vector, and performing quantitatively better on these abilities on various retrieval metrics than classical trained on discriminative tasks in the zero-shot setting. We find that the primary gains are powerful zero-shot priors, and that rendering-based and classic models both perform well once the model is given access to database images at training time. As well as achieving improved performance on 2D image retrieval, this opens up new possibilities for non-generative tasks, including 3D decision-making, recognition, and visual question-answering, and suggests that a promising direction forward in those areas involves adapting the priors of existing state-of-the-art generative models.
8 Acknowledgements
The authors would like thank Fangyin Wei, Professor Felix Heide, Evan Dogariu, and other students in COS526 for helpful discussions and kind support.
References
- [1] Aishwarya Agrawal, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C. Lawrence Zitnick, Dhruv Batra, and Devi Parikh. Vqa: Visual question answering, 2016.
- [2] Karim Amer, Mohamed Samy, Mahmoud Shaker, and Mohamed ElHelw. Deep convolutional neural network based autonomous drone navigation. In Thirteenth International Conference on Machine Vision, volume 11605, pages 16–24. SPIE, 2021.
- [3] Angel X. Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, Jianxiong Xiao, Li Yi, and Fisher Yu. Shapenet: An information-rich 3d model repository, 2015.
- [4] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations, 2020.
- [5] Ting Chen, Lala Li, Saurabh Saxena, Geoffrey Hinton, and David J Fleet. A generalist framework for panoptic segmentation of images and videos. arXiv preprint arXiv:2210.06366, 2022.
- [6] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255, 2009.
- [7] Jiankang Deng, Jia Guo, Niannan Xue, and Stefanos Zafeiriou. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4690–4699, 2019.
- [8] Yiqun Duan, Xianda Guo, and Zhengbiao Zhu. Diffusiondepth: Diffusion denoising approach for monocular depth estimation. ArXiv, abs/2303.05021, 2023.
- [9] Zhangxuan Gu, Haoxing Chen, Zhuoer Xu, Jun Lan, Changhua Meng, and Weiqiang Wang. Diffusioninst: Diffusion model for instance segmentation. ArXiv, abs/2212.02773, 2022.
- [10] Heewoo Jun and Alex Nichol. Shap-e: Generating conditional 3d implicit functions, 2023.
- [11] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009.
- [12] Axel Levy, Mark Matthews, Matan Sela, Gordon Wetzstein, and Dmitry Lagun. Melon: Nerf with unposed images using equivalence class estimation, 2023.
- [13] Ruoshi Liu, Rundi Wu, Basile Van Hoorick, Pavel Tokmakov, Sergey Zakharov, and Carl Vondrick. Zero-1-to-3: Zero-shot one image to 3d object, 2023.
- [14] Weiyang Liu, Yandong Wen, Zhiding Yu, Ming Li, Bhiksha Raj, and Le Song. Sphereface: Deep hypersphere embedding for face recognition, 2018.
- [15] Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis, 2020.
- [16] Thomas Müller, Alex Evans, Christoph Schied, and Alexander Keller. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph., 41(4):102:1–102:15, July 2022.
- [17] Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. Zero-shot text-to-image generation, 2021.
- [18] Florian Schroff, Dmitry Kalenichenko, and James Philbin. FaceNet: A unified embedding for face recognition and clustering. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, jun 2015.
- [19] Thalles Santos Silva. Exploring simclr: A simple framework for contrastive learning of visual representations. https://sthalles.github.io, 2020.
- [20] Mingxing Tan and Quoc V. Le. Efficientnet: Rethinking model scaling for convolutional neural networks, 2020.
- [21] Yandong Wen, Kaipeng Zhang, Zhifeng Li, and Yu Qiao. A discriminative feature learning approach for deep face recognition. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision – ECCV 2016, pages 499–515, Cham, 2016. Springer International Publishing.
- [22] Lilian Weng. Contrastive representation learning. lilianweng.github.io, May 2021.
- [23] Julia Wolleb, Robin Sandkühler, Florentin Bieder, Philippe Valmaggia, and Philippe C Cattin. Diffusion models for implicit image segmentation ensembles. In International Conference on Medical Imaging with Deep Learning, pages 1336–1348. PMLR, 2022.
- [24] Junde Wu, Huihui Fang, Yu Zhang, Yehui Yang, and Yanwu Xu. Medsegdiff: Medical image segmentation with diffusion probabilistic model. arXiv preprint arXiv:2211.00611, 2022.
- [25] Alex Yu, Vickie Ye, Matthew Tancik, and Angjoo Kanazawa. pixelnerf: Neural radiance fields from one or few images, 2021.







