Local Spatiotemporal Representation Learning for Longitudinally-consistent Neuroimage Analysis
Abstract
Recent self-supervised advances in medical computer vision exploit the global and local anatomical self-similarity for pretraining prior to downstream tasks such as segmentation. However, current methods assume i.i.d. image acquisition, which is invalid in clinical study designs where follow-up longitudinal scans track subject-specific temporal changes. Further, existing self-supervised methods for medically-relevant image-to-image architectures exploit only spatial or temporal self-similarity and do so via a loss applied only at a single image-scale, with naive multi-scale spatiotemporal extensions collapsing to degenerate solutions. To these ends, this paper makes two contributions: (1) It presents a local and multi-scale spatiotemporal representation learning method for image-to-image architectures trained on longitudinal images. It exploits the spatiotemporal self-similarity of learned multi-scale intra-subject image features for pretraining and develops several feature-wise regularizations that avoid degenerate representations; (2) During finetuning, it proposes a surprisingly simple self-supervised segmentation consistency regularization to exploit intra-subject correlation. Benchmarked across various segmentation tasks, the proposed framework outperforms both well-tuned randomly-initialized baselines and current self-supervised techniques designed for both i.i.d. and longitudinal datasets. These improvements are demonstrated across both longitudinal neurodegenerative adult MRI and developing infant brain MRI and yield both higher performance and longitudinal consistency.
1 Introduction
Tracking subject-specific anatomical trends over time is crucial to both clinical diagnostics and large-scale biomedical science. Such longitudinal imaging is especially relevant to analyzing neurological patterns of growth and degeneration via brain imaging in pediatric and elderly populations, respectively. As tracking individual structural changes requires precise and longitudinally-consistent segmentation methods with scarce annotated training volumes, we identify two major bottlenecks in existing self-supervised biomedical image analysis methods which we use to motivate our work.
Learning with few annotations. While modern imaging studies may scan hundreds to thousands of individuals, manually outlining volumetric structures of interest across multiple individuals for supervised segmentation training is prohibitively expensive. Therefore, current work focuses on leveraging large sets of unlabeled images to pretrain image-to-image architectures (e.g., the U-Net [48]), which can then be efficiently finetuned in the one or few-shot setting. These self-supervised methods may handcraft pre-text training objectives [10, 19, 45, 65] or may attempt to pretrain the base network to be equivariant to transformations in order to preserve semantic meaning as in contrastive learning in medical [9, 63] and natural image vision [2, 25, 38, 60, 67, 68]. However, these methods typically leverage label supervision in sampling for their losses and impose a self-supervised loss only at a single feature scale (typically the encoder bottleneck or network output). Naive application of local unsupervised multi-scale contrastive losses [16, 44] lead to degenerate representations (Fig. 1, row A) when applied to both the encoder and decoder of image-to-image architectures.

Violating i.i.d. assumptions. Further, most existing self-supervised frameworks assume i.i.d. data. Unfortunately, this assumption does not transfer to longitudinal studies where intra-subject temporal images are highly correlated. Emerging longitudinal representation learning methods focus on imposing temporal-consistency into the encoder bottleneck [15, 43, 66], such that the encoder learns representations that are aware of the order of acquisition [15] and the overall trajectory [43]. These methods address image-level tasks such as disease classification or progression and age prediction. However, their extension to pixel-level applications with image-to-image architectures remains unclear.
Methods. Motivated by the above limitations, in this work, we claim that the spatiotemporal dependency of imaging data should be explicitly incorporated in self-supervised frameworks. We do so by exploiting the spatial and temporal self-similarity of local multiscale deep features in both encoder and decoder and further learn diverse intermediate representations by developing regularizations for self-supervised similarity objectives. Lastly, when finetuning with limited annotated data, we encourage predictions on unlabeled subject-wise images to be spatiotemporally consistent.
Contributions. This work makes the following contributions: (1) It presents a longitudinally-consistent spatiotemporal representation learning framework to learn from image time-series; (2) To impose multi-scale local self-supervision while avoiding degenerate solutions, it develops regularization terms on the variance, covariance, orthogonality of local features within the decoder; (3) To further self-supervise the fine-tuning stage, the proposed method encourages segmentations from adjacent timepoints on unlabeled data to be consistent; (4) Across three large-scale longitudinal one, few, and full-shot segmentation tasks on both elderly and pediatric populations, the developed framework yields improved segmentation performance and higher longitudinal segmentation consistency. Our code is available at https://www.mengweiren.com/research/spatiotemporal-learning/.
2 Related work
Self-supervision. Self-supervised learning (SSL) methods aim to learn hierarchical representations from unannotated data, which can then be transferred to tasks operating in low-annotation regimes. Early work focused on pretext tasks where a handcrafted loss is used to pretrain networks via orientation prediction [19], context restoration [10, 45], channel prediction [65], among others. Given their heuristic nature and suboptimal generalization, recent work instead focuses on data-driven SSL losses.
Contrastive learning. Constrastive learning [11, 22, 23, 24, 49, 55] (CL) typically transforms an input image and asks the embeddings of the input image and its transformation (the positive pair) to be close to one another and far apart from embeddings of other images (negatives) via a noise contrastive estimation [21, 42] (NCE) loss. While performant on image-level recognition tasks [12, 32], CL requires non-trivial modification to extend to pixel-level segmentation tasks, as described below.
Negative-free representation learning. In several applications, true negative samples may be difficult to construct [26]. For example, when learning on intra-domain internal image patches [16, 44], non-local spatial positions may be semantically similar, but NCE objectives push their embeddings apart, leading to false negative pairs introducing label noise in the training objective. This drawback may be mitigated via data-driven SSL methods which only use positive samples and avoid low-diversity (or collapsed [30]) embeddings solutions via predictor networks and custom backpropogation [13, 20] and careful regularization [6, 62]. However, as above, these methods operate on global image embeddings and require modification for pixel-level tasks.
Spatial self-supervision. Towards downstream segmentation, recent work [2, 25, 35, 38, 60, 67, 68] encourages local single-scale features either within an image or across images to cluster semantically by constructing positive pairs using ground-truth labels. To incorporate local and multi-scale spatial considerations into unpaired image translation and registration [16, 44] imposed contrastive losses on randomly-sampled layer-wise encoder features by considering corresponding spatial indices as positives and all other locations as negatives. Our work builds on this by instead only considering temporal positives (described below) in the layerwise losses alongside custom regularization which avoids low-diversity decoder embeddings observed with naive application in Fig. 1.
Temporal self-supervision. SSL methods developed for video achieve high performance by exploiting temporally consistent transformation [5] and temporal pretext task [14, 29, 59]. However, longitudinal biomedical image time-series have sparser sampling (typically 2–5 timepoints/subject) and have greater spatial extents (volumes instead of images), which leads to distinct modeling considerations.
Emerging biomedical methods [15, 43, 66] enforce smooth trajectories for subject-wise images in the encoder latent space and deploy their methods on image-level downstream tasks such as disease classification and age regression. However, these methods focus on learning a global embedding, without a clear extension to pixel-level tasks such as segmentation.
Biomedical image segmentation. Major challenges specific to biomedical segmentation include: (1) large 3D volumes; (2) limited sample sizes and annotations; and (3) non-i.i.d. longitudinal acquisitions tracking temporal anatomical changes. To these ends, conventional approaches use a combination of intensity-based probabilistic models and registration-driven atlas-based models [1, 27, 39, 58]. In particular, longitudinal image analysis typically makes use of one or few longitudinal atlases [28, 33, 47, 51, 52], which motivates the one and few-shot segmentation settings benchmarked in this paper, respectively.
More recently, deep segmentation networks achieve strong performance [8, 7, 40, 41, 48] given enough training volumes. In the low-annotation setting, weakly supervised methods develop custom loss functions [31, 37], but may have drawbracks analogous to the handcrafted SSL losses described above. Fortunately, recent data-driven self and semi-supervised methods are well-suited to pixel-level prediction. For example, to pretrain an encoder for segmentation, [9, 63] develop application-specific positive and negative sampling strategies for contrastive training, where 2D slices from similar locations in registered 3D volumes across subjects constitute positive pairs. While these methods have been successful in their applications, they are inherently slice-based methods and are outperformed by well-tuned randomly-initialized 3D baselines on our datasets (Tab. 1). Further, these self-supervised biomedical segmentation methods do not explicitly account for non i.i.d acquisitions. Lastly, to our knowledge, existing longitudinal deep learning work developed for biomedical segmentation is currently very specific to its target application [18, 36, 56] (for example, in tasks such as MS lesion change detection [56]) or require supervised pre-training on annotated cross-sectional datasets [61], whereas we develop a generic self-supervised spatiotemporal representation learning framework for non-i.i.d. longitudinal data which can be applied to any downstream task in principle.
3 Methodology
Fig. 2 illustrates the proposed framework. The base U-Net architecture is pretrained end-to-end in a self-supervised manner with intra-subject spatiotemporal losses. On convergence, the pretrained network parameters serve as an initialization for training downstream local spatiotemporal pixel-level tasks (e.g., registration or segmentation). This work will focus on downstream segmentation in the one, few, and full-shot regimes. The main similarity loss is applied on multiscale local patches from different timepoints of the same subject, which attracts features in corresponding locations in separate timepoints together. The high-dimensional U-Net features corresponding to the boxes at the same locations in Fig. 2a should maintain high similarity despite varying appearance. The feature regularizers (Fig. 2b,c,d) avoid degenerate U-Net decoder embeddings in the patch similarity training and the output regularizer (Fig. 2e) encourages finetuning consistency on unannotated data.

Setup. The unlabeled dataset is a collection of subjects, where each subject has at least two longitudinal image acquisitions available during pretraining. During every iteration, a pair of images are randomly sampled from subject at distinct timepoints and , where . indicates the number of registered images from subject , and is a 3D volume of spatial dimension and channels. These channels are typically multi-modality acquisitions (e.g., T1w and T2w MRI from the same subject).
Spatiotemporal patchwise similarity loss. We aim to associate the local embeddings of and such that the representations are longitudinally-aware. To this end, a weight-sharing 3D U-Net takes both images as input and produce a set of multi-scale CNN features , where each element indicates the output of the th layer of interest. feature vectors are then randomly sampled from the 3D spatial indices of the feature map, where each feature vector represents a local patch of the input image. These patch-wise activations from multiple layers of form hierarchical representations of local regions of the input image.
To maximize the similarity of corresponding local features (e.g., blue boxes in Fig. 2a) without negative samples, we use a projector MLP and a predictor head and extend [13] to patchwise operation such that matching spatial indices have high agreement. The patchwise similarity loss between two representations is defined as follows: where , and . The total loss is an average of all sampled patches, across multilayer features:
(1) |
Architectural challenges in multi-scale representation learning. applied to the hidden layers of a U-Net is found to maximize patchwise similarity of the encoder features (as expected) but lead to low-diversity and semantically-incoherent representations in the decoder layers as observed from the similarity maps in Fig. 1B and empirical observations in App. E. We speculate that this is partially attributable to U-Net skip connections, which lead to a degenerate solution where the encoder learns representations which are good enough to minimize the decoder losses and the decoder layers do not have to learn useful representations which transfer. To this end, we develop several regularization strategies such that the decoder layers obtain diverse and semantically-coherent embeddings.
Orthogonality. During model prototyping, we empirically observe that low-diversity embeddings first originate in the U-Net bottleneck (Fig. 1B cols 4,5 and App. Fig. 14 rows 2,3) which are then upsampled hierarchically through the decoder. We therefore encourage decoupled bottleneck features between the encoder and decoder. Revisiting the U-Net skip-connection, the encoder features from the -th layer are concatenated with the upsampled decoder features , followed by a convolution to attain the same feature dimension as . This yields and by regularizing for orthogonality between and in the projector embedding space, we implicitly encourage to learn better representations instead of converging to degenerate solutions. The orthogonality loss is defined as
(2) |
where and are projected representations via at sampling location , from the matched encoder/decoder layers, respectively.
Variance and covariance. We further encourage spatial variation and channel-wise decorrelation of local decoder features to avoid degenerate representations. We extend the variance and covariance terms of [6] towards patchwise multi-layer operation and apply them to the decoder layers. A standard deviation loss encourages spatial feature variation above a threshold of and is defined as
(3) |
where is the standard deviation of randomly-sampled projected features from the -th layer and is added for numerical stability. A covariance regularization decorrelates channelwise activations in the decoder to prevent low-diversity embeddings by minimizing
(4) |
where is the covariance matrix of -D representation and indicates its off-diagonal indices.
Reconstruction. To further pretrain the decoder, we investigate reconstruction losses commonly used in unsupervised learning. However, as high-resolution information is passed through skip connections, a reconstruction loss for a U-Net is near-trivially minimized and does not address degenerate solutions. Therefore, inspired by denoising autoencoders [57], we encourage network equivariance and invariance to geometric () and intensity-based transformations (), respectively, of the input image , by modifying the reconstruction objective to a denoising loss as,
(5) |
Finetuning with longitudinal consistency regularization. In longitudinal segmentation, if the finetuning data do not cover the overall age range, the network may perform well on the finetuned timepoints, but may perform poorly on unseen ages even when pretrained on unannotated data across all ages. We therefore develop a self-supervised longitudinal consistency regularization term applied at the network output during finetuning to increase the intra-subject agreement. Given registered and unannotated image volumes, we formulate a segmentation prediction consistency loss as,
(6) |
which is minimized alongside the supervised segmentation term below during finetuning.
Total objective. During pretraining, our overall objective function is a weighted sum of the patch similarity loss, reconstruction loss, and three regularizations, and is defined as,
(7) |
During finetuning, we use a combined dice and cross entropy loss on supervised training pairs as,
(8) |
where is the groundtruth segmentation label. Lastly, we additionally use the segmentation consistency loss for non i.i.d inputs, which forms our final finetuning loss .

4 Experiments
Data and segmentation tasks. We conduct experiments on two de-identified longitudinal neuroimaging datasets, and specifically design three tasks to benchmark different extents of biomedical domain gaps between the finetuning and testing data. The main body of this work focuses on one-shot segmentation using one annotated subject (a common medical image analysis setting). Benchmarks on few-shot and fully-supervised segmentation tasks are provided in Appendix A. For both datasets, we perform a train/validation/test split on a subject-wise basis with 70%, 10% and 20% of the participants. The validation set is used for model and hyperparameter selection and results are reported on a held-out test set. ANTs [3, 4] is used to perform inter-subject affine alignment for all experiments, followed by intra-subject deformable registration to obtain accurate spatiotemporal correspondence for and calculation. All images are skull-stripped, bias-field corrected, and intensity-normalized. Further data preprocessing and splitting details are described in Appendix C.
OASIS3 [34] is a publicly-available dataset consisting of 1639 brain MRI scans of 992 longitudinally imaged subjects. Each subject has 1–5 temporal acquisitions over a -year long observation window, resulting in an aging cohort over the span 42 to 95 years which includes cognitively normal and mildly impaired individuals alongside subjects with Alzheimer’s Disease. On OASIS3, we tackle whole-brain segmentation using the FreeSurfer label convention [17]. Cross-sectional FreeSurfer anatomical segmentation was done as part of the data release. Observing strong temporal inconsistency (see App. C.2), we further perform longitudinal FreeSurfer [47] to improve the temporal consistency of the reference segmentation. We exclude labels that have less than 100 voxels in all subjects, which results in 33 labels for segmentation training and evaluation. Finetuning for one-shot segmentation is performed on a single FreeSurfer-annotated subject with four timepoints.
IBIS is an infant brain imaging study, which longitudinally acquires 1272 structural T1w/T2w MRI from 552 infants across both controls and infants at a high-risk for Autism Spectrum Disorder (ASD) over a span of 3 to 36 months of age. We tackle two distinct tasks: subcortical segmentation (IBIS-subcort) and white/gray matter tissue segmentation (IBIS-wmgm). For IBIS-subcort, a multi-atlas method [54] cross-sectionally segments sub-cortical grey matter (relevant to ASD [49]) into 13 structures of interest, which are then followed by manual corrections. For one-shot benchmarking, finetuning is performed only on a single longitudinally-labeled subject, similar to OASIS3 above. Further, we use the IBIS-wmgm setting to simulate a real-world use-case detailed in App. D.2. Briefly, brain MRI segmentation into grey/white matter is straightforward at 24-36 months of age due to the presence of anatomical edges in the images. However, 6-month-old grey/white matter brain segmentation remains elusive without manually labeled datasets for supervision due to white matter myelination leading to isointense appearance at that age [53] (e.g., Fig. 2a and App. D.2). We therefore investigate finetuning all benchmarked methods on a single 36 month old image which can be reliably segmented with [46] and then evaluate segmentation deployment with a strong domain shift on 6 month old isointense images and labels (whose ground truth labels are generated by a fully supervised external model [64]).
Baselines and Evaluation Strategies. We analyze segmentation performance and longitudinal consistency against well-tuned randomly initialized 2D/3D U-Nets (RandInitUnet.2D/3D) and various high-performing self-supervised pretraining methods. These include the pretext-task based Context Restoration [10], longitudinal representation learning based LNE [43], along with the contrastive learning based GLCL [9] and PCL [63] methods which operate on image slices. We also repurpose PatchNCE [44] for segmentation to evaluate its generic representation learning capabilities. All methods are pretrained and finetuned with both geometric and intensity-based augmentation, and share the same network architecture.
We quantify network performance via commonly used scores such as the Dice coefficient, IoU, and the 95-th percentile of the Hausdorff distance. More importantly, we also quantify the longitudinal agreement between intra-subject non-linearly registered temporal segmentations via scores such as the spatiotemporal consistency of segmentation [36] where and are temporal segmentation predictions from non-linearly registered input images; and the absolute symmetrized percent change [47] , where and are the volume of a structure calculated from and . We also report and on the groundtruth segmentations as a reference.
Implementation details. We train a 3D U-Net [48] as the base image-to-image architecture with four levels of up/down sampling and repeated Conv-BN-ReLU blocks (all architectural details are provided in App. B). The projector head consists of a 3-layer MLP with 2048 nodes per layer. Following [13], we apply batch normalization after each MLP, followed by ReLU activation, and normalization of the final activation. The predictor is a 3-layer bottlenecked MLP with widths of 2048-256-2048. We apply geometric (left-right flip, random affine warps) and intensity augmentations including random blur, noise, gamma contrast enhancement. Additional MRI-specific augmentations with random bias field and motion artifacts are also applied and are followed by random spatial cropping. We use a batch size of 3 crops and an initial learning rate of for both pretraining and finetuning. All networks are trained with the Adam optimizer ( during pretraining and during finetuning and in both settings) on a single Nvidia RTX8000 GPU (45GB vRAM). The networks are pretrained for a maximum of steps and the best model based on validation performance is used for fine-tuning for another steps, alongside linear learning rate decay. All experiments are run on a fixed random seed due to limited computational budgets. Based on the ablation analysis in Tab. 2, we empirically choose for all datasets, and use for OASIS3, for IBIS. Further details on the configurations, implementation of our method and other baselines are provided in Appendix B.
Method | IBIS-subcort | IBIS-wmgm | OASIS3 | ||||||
---|---|---|---|---|---|---|---|---|---|
IoU | HD95 | ASPC | IoU | HD95 | ASPC | IoU | HD95 | ASPC | |
GT | - | - | 7.819 | - | - | - | - | - | 3.947 |
RandInitUnet.2D | 0.707 | 4.485 | 7.127 | 0.510 | 3.274 | 11.506 | 0.687 | 2.545 | 10.189 |
RandInitUnet.3D | 0.720 | 2.892 | 4.644 | 0.560 | 3.788 | 5.515 | 0.715 | 2.206 | 2.789 |
Context Restore [10] | 0.711 | 4.403 | 7.831 | 0.444 | 8.273 | 29.235 | 0.717 | 3.323 | 5.577 |
LNE [43] | 0.736 | 3.033 | 5.866 | 0.563 | 3.201 | 5.352 | 0.726 | 1.988 | 8.836 |
GLCL [9] | 0.718 | 3.203 | 5.514 | 0.550 | 4.112 | 8.472 | 0.695 | 2.264 | 4.622 |
PCL [63] | 0.713 | 3.270 | 5.610 | 0.562 | 4.974 | 10.648 | 0.707 | 2.327 | 4.850 |
PatchNCE [44] | 0.743 | 1.266 | 5.780 | 0.607 | 4.344 | 3.782 | 0.738 | 2.275 | 3.114 |
Ours w/o | 0.754 | 1.145 | 5.483 | 0.614 | 3.291 | 2.462 | 0.739 | 1.940 | 2.729 |
Ours w/ | 0.757 | 1.178 | 4.475 | 0.676 | 3.237 | 4.155 | 0.737 | 2.094 | 2.754 |

Segmentation and longitudinal consistency results. Fig. 3 qualitatively demonstrates improved generalization using our method on unseen longitudinal data (row 1-3), especially on data displaying rapid intra-subject temporal developments (IBIS-wmgm,subcort). These improvements are consistent with the quantitative results presented in Fig. 4/Tab. 1 which indicate both improved segmentation performance and longitudinal consistency. In the strong domain shift setting of IBIS-wmgm, we see a near ten-point increase in median dice over most baselines. With moderate shifts in IBIS-subcort, we see appreciable increases in performance and consistency. We note that brain segmentation on adult brain MRI data from OASIS3 is a comparatively easier task as adult neuroimages do not significantly change appearance between imaging sessions. Therefore, several baselines are able to match (but not exceed) the segmentation performance of our method on OASIS3. However, all baselines are outperformed by ours on all datasets in terms of longitudinal-consistency which is essential to non-i.i.d. statistical analysis. In particular, both our pretraining (Ours w/o ) and finetuning (Ours w/ ) methods show STCS and ASPC improvements over all of the compared settings. Fig. 3 (bottom row) shows an example of temporal predictions on two unseen timepoints from IBIS-wmgm. The predictions from Context Restoration [10] match only the input image intensity and lack anatomical and longitudinal consistency, LNE [43] introduces false positive predictions in temporal lobe (within the orange circle), and the proposed method yields a more spatiotemporally and anatomically consistent segmentation. Beyond one-shot segmentation, we also observe gains in the few-shot and fully-supervised segmentation settings in Appendix A and Suppl. Tabs. 3 and 4, respectively.
Qualitative self-supervised spatiotemporal similarity. In Fig. 1, we qualitatively examine the learned visual representations of the proposed method via intra-subject temporal self-similarity (C) and compare it to two of its variants which either use contrastive learning with unsupervised negatives (A) or negative-free representation learning (B). We calculate the per-layer multiscale feature self-similarity between the query and each key from the intra-subject feature maps at a different age (blue box). In row A, we see that assuming that all spatial indices not in correspondence constitute negative pairs leads to highly-positionally dependent representations in the decoder which carry low semantic meaning (e.g., in the adult data, the similarity to localize to the ventricles in the coronal view). By discarding all negatives in row B, we observe semantically-incoherent and low-diversity embeddings and artifacts in the decoder layers on both datasets. Finally, with careful regularization in row C, our methods discards all negative pairs and attains semantically and positionally relevant representations.
Ablations. As the proposed method consists of several moving parts, an ablation analysis is conducted over different model configurations, hyperparameters, and loss functions, reported in Tab. 2 consisting of average dice coefficients. The combination of all proposed components yields optimal results. Further ablations and baseline tuning results are reported in Appendix A.
Row A starts with a base setting where only four encoder layers from the U-Net are selected for computation and a small MLP width of 256 is used for the projection and prediction heads. Here, IBIS-subcort and OASIS3 results are competitive with randomly initialized U-Net, as expected given the lack of auxiliary losses, data augmentation, and regularizations. However, the IBIS-wmgm experiment already shows a 2% improvement over random initialization, indicating benefits of using patchwise similarity losses for better out-of-distribution generalization even with suboptimal setups.
Row B: With larger projector and predictor networks, we observe improvements on two out of three datasets, which is consistent with trends observed on natural images [12, 13].
Rows C–F: On adding decoder layers to and introducing alongside data augmentation (without any regularization), we typically observe inconsistent dataset-specific trends which arise from unregularized representations (e.g., Fig. 1B). We speculate that a poorly trained decoder (due to a lack of regularization) may be equivalent to random initialization in the context of pretraining for segmentation tasks. However, a combination of these components (Row F) leads to an appreciable increase in performance.
Rows G–K: When orthogonal regularization and/or covariance/variance regularization is used, we observe the best performance when they are applied together alongside augmentation and . In rows J and K, we observe that different hyperparameters are optimal for OASIS3 and IBIS (which already outperform all baseline methods in Tab. 1), which is intuitive as these are drastically different cohorts.
Row L: Finally, the overall proposed model is achieved when is added to the finetuning objective, which yields strong improvements for IBIS-{wmgm,subcort} and maintains OASIS3 performance.
Exp | Loss Layers | #MLP | Aug. | IBIS-subcort | IBIS-wmgm | OASIS3 | |||||
---|---|---|---|---|---|---|---|---|---|---|---|
A | Enc | 256 | 0.829(0.068) | 0.733(0.062) | 0.783(0.16) | ||||||
B | Enc | 2048 | 0.849(0.060) | 0.732(0.073) | 0.809(0.13) | ||||||
C | Enc | 2048 | 0.859(0.058) | 0.713(0.079) | 0.811(0.13) | ||||||
D | EncDec | 2048 | 0.860(0.058) | 0.718(0.066) | 0.810(0.13) | ||||||
E | EncDec | 2048 | 0.858(0.060) | 0.724(0.077) | 0.809(0.13) | ||||||
F | EncDec | 2048 | 0.856(0.060) | 0.739(0.067) | 0.812(0.13) | ||||||
G | EncDec | 2048 | 100 | 0.857(0.062) | 0.728(0.074) | 0.809(0.13) | |||||
H | EncDec | 2048 | 0.845(0.063) | 0.739(0.074) | 0.804(0.13) | ||||||
I | EncDec | 2048 | 100 | 0.859(0.056) | 0.735(0.061) | 0.811(0.13) | |||||
J | EncDec | 2048 | 100 | 0.863(0.057) | 0.758(0.062) | 0.808(0.14) | |||||
K | EncDec | 2048 | 100 | 0.853(0.055) | 0.745(0.058) | 0.813(0.13) | |||||
L | EncDec | 2048 | 100 | * | 0.870(0.052) | 0.806(0.030) | 0.810(0.13) |
5 Discussion
Limitations and future work. The presented work opens up many follow-up questions which will be tackled in future work: (1) Our proposed losses enable better training and performance, but require the tuning of several regularization weights and layer selections which may reveal dataset specific patterns (e.g., rows J, K in Tab. 2). The weights and layers selected here were chosen based on limited exploratory experiments on the validation sets due to computational budgets and future work will exhaustively search the hyperparameter space for optimal performance. (2) Our pretraining assumes accurate non-linear intra-subject registration, which may be non-trivial in edge cases like modalities with strong distortion and artifacts (e.g., eddy corruption in diffusion MRI). However, when studying pre and post-operative imaging (e.g., surgical excision of lesions), large topological changes break the assumptions of our model and will require the development of lesion-masked positive patch sampling methods. (3) We use a two-stage pre-training and finetuning approach and it is plausible that the proposed method can be reduced to a single-stage combined framework. (4) While this paper focused on downstream segmentation finetuning, the pretraining framework is generic to any pixel-level task (e.g., registration) and its extension to such tasks will be explored. (5) While the proposed methods yield strong longitudinal segmentation consistency improvements across all datasets, we note that absolute segmentation performance gains in an elderly cohort (OASIS3) are modest in comparison to the rapidly developing infant dataset (IBIS) where higher gains are achieved. Future work will further investigate these performance differences between populations of differing temporal trends. (6) We extended the negative-free framework of [13] to patchwise operations for its relative simplicity and it is plausible that other negative-free similarity terms [6, 20, 62] may further improve results. (7) In our data preparation for pretraining, subject-wise image time-series were registered to a single time-point instead of a subject-specific template, which is known to increase statistical bias [47].
Limited scope. The proposed method generically applies to medical image time-series and we do not anticipate negative impacts beyond those that currently exist for segmentation methods. However, while our tasks have impacts on understanding real-world disease mechanisms, we cannot claim any further insight into differences between subpopulations, as such analysis requires close collaboration with clinicians, neuroscientists, and biostatisticians, which is beyond the scope of this work.
Conclusions. This paper addressed several open questions regarding the self-supervised pretraining and finetuning of image-to-image architectures on longitudinal volumes using objective functions which exploit both intra-subject spatial and temporal self-similarity. It developed a local negative sample-free framework that trains multiple multi-scale hidden layers of image-to-image architectures that then enabled improved downstream segmentation performance, all while achieving semantically-meaningful representations via careful regularization of the decoder activations. During finetuning, it similarly developed a simple consistency-regularization objective which encourages longitudinal agreement between predictions on unlabeled data. When applied to large-scale neurodeveloping and neurodegerative longitudinal images, the proposed framework yielded improved segmentation performance and temporal consistency, both of which are crucial to statistical analyses of mechanisms of interest such as Alzheimer’s Disease (OASIS3) and Autism Spectrum Disorder (IBIS).
Acknowledgements
The authors are grateful to NIH R01-HD055741-12, R01-MH118362-02S1, 1R01MH118362-01, 1R01HD088125-01A1, R01MH122447, R01-HD059854, U54-HD079124, P50-HD103573, R01ES032294, and the NYS Center for Advanced Technology in Telecommunications (CATT).
Adult longitudinal T1w MRI data were provided by OASIS-3; NIH P50 AG00561, P30 NS09857781, P01 AG026276, P01 AG003991, R01 AG043434, UL1 TR000448, R01 EB009352. Longitudinal developing infant brain T1w/T2w MRI were provided by the Infant Brain Imaging Study (IBIS) Network, which is an NIH-funded Autism Centers of Excellence (ACE) project and consists of a consortium of 10 universities in the U.S. and Canada.
References
- [1] Paul Aljabar, Rolf A Heckemann, Alexander Hammers, Joseph V Hajnal, and Daniel Rueckert. Multi-atlas based segmentation of brain images: atlas selection and its effect on accuracy. Neuroimage, 46(3):726–738, 2009.
- [2] Iñigo Alonso, Alberto Sabater, David Ferstl, Luis Montesano, and Ana C. Murillo. Semi-supervised semantic segmentation with pixel-level contrastive learning from a class-wise memory bank. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 8219–8228, October 2021.
- [3] Brian B Avants, Charles L Epstein, Murray Grossman, and James C Gee. Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain. Medical image analysis, 12(1):26–41, 2008.
- [4] Brian B Avants, Paul Yushkevich, John Pluta, David Minkoff, Marc Korczykowski, John Detre, and James C Gee. The optimal template effect in hippocampus studies of diseased populations. Neuroimage, 49(3):2457–2466, 2010.
- [5] Yutong Bai, Haoqi Fan, Ishan Misra, Ganesh Venkatesh, Yongyi Lu, Yuyin Zhou, Qihang Yu, Vikas Chandra, and Alan Yuille. Can temporal information help with contrastive self-supervised learning? arXiv preprint arXiv:2011.13046, 2020.
- [6] Adrien Bardes, Jean Ponce, and Yann LeCun. Vicreg: Variance-invariance-covariance regularization for self-supervised learning. arXiv preprint arXiv:2105.04906, 2021.
- [7] Benjamin Billot, Magdamo Colin, Sean E Arnold, Sudeshna Das, Juan Iglesias, et al. Robust segmentation of brain mri in the wild with hierarchical cnns and no retraining. arXiv preprint arXiv:2203.01969, 2022.
- [8] Benjamin Billot, Douglas N Greve, Oula Puonti, Axel Thielscher, Koen Van Leemput, Bruce Fischl, Adrian V Dalca, and Juan Eugenio Iglesias. Synthseg: Domain randomisation for segmentation of brain mri scans of any contrast and resolution. arXiv preprint arXiv:2107.09559, 2021.
- [9] Krishna Chaitanya, Ertunc Erdil, Neerav Karani, and Ender Konukoglu. Contrastive learning of global and local features for medical image segmentation with limited annotations. Advances in Neural Information Processing Systems, 33:12546–12558, 2020.
- [10] Liang Chen, Paul Bentley, Kensaku Mori, Kazunari Misawa, Michitaka Fujiwara, and Daniel Rueckert. Self-supervised learning for medical image analysis using image context restoration. Medical image analysis, 58:101539, 2019.
- [11] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597–1607. PMLR, 2020.
- [12] Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey E Hinton. Big self-supervised models are strong semi-supervised learners. Advances in neural information processing systems, 33:22243–22255, 2020.
- [13] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15750–15758, June 2021.
- [14] Hyeon Cho, Taehoon Kim, Hyung Jin Chang, and Wonjun Hwang. Self-supervised spatio-temporal representation learning using variable playback speed prediction. arXiv preprint arXiv:2003.02692, 8, 2020.
- [15] Raphaël Couronné, Paul Vernhet, and Stanley Durrleman. Longitudinal self-supervision to disentangle inter-patient variability from disease progression. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 231–241. Springer, 2021.
- [16] Neel Dey, Jo Schlemper, Seyed Sadegh Mohseni Salehi, Bo Zhou, Guido Gerig, and Michal Sofka. Contrareg: Contrastive learning of multi-modality unsupervised deformable image registration. In Medical Image Computing and Computer Assisted Intervention – MICCAI 2022. Springer, 2022.
- [17] Bruce Fischl, David H Salat, Evelina Busa, Marilyn Albert, Megan Dieterich, Christian Haselgrove, Andre Van Der Kouwe, Ron Killiany, David Kennedy, Shuna Klaveness, et al. Whole brain segmentation: automated labeling of neuroanatomical structures in the human brain. Neuron, 33(3):341–355, 2002.
- [18] Yang Gao, Jeff M Phillips, Yan Zheng, Renqiang Min, P Thomas Fletcher, and Guido Gerig. Fully convolutional structured lstm networks for joint 4d medical image segmentation. In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pages 1104–1108. IEEE, 2018.
- [19] Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised representation learning by predicting image rotations. arXiv preprint arXiv:1803.07728, 2018.
- [20] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems, 33:21271–21284, 2020.
- [21] Michael Gutmann and Aapo Hyvärinen. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 297–304. JMLR Workshop and Conference Proceedings, 2010.
- [22] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9729–9738, 2020.
- [23] Olivier Henaff. Data-efficient image recognition with contrastive predictive coding. In International Conference on Machine Learning, pages 4182–4192. PMLR, 2020.
- [24] R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. arXiv preprint arXiv:1808.06670, 2018.
- [25] Hanzhe Hu, Jinshi Cui, and Liwei Wang. Region-aware contrastive learning for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 16291–16301, October 2021.
- [26] Tri Huynh, Simon Kornblith, Matthew R Walter, Michael Maire, and Maryam Khademi. Boosting contrastive self-supervised learning with false negative cancellation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 2785–2795, 2022.
- [27] Juan Eugenio Iglesias and Mert R Sabuncu. Multi-atlas segmentation of biomedical images: a survey. Medical image analysis, 24(1):205–219, 2015.
- [28] Juan Eugenio Iglesias, Koen Van Leemput, Jean Augustinack, Ricardo Insausti, Bruce Fischl, Martin Reuter, Alzheimer’s Disease Neuroimaging Initiative, et al. Bayesian longitudinal segmentation of hippocampal substructures in brain mri using subject-specific atlases. Neuroimage, 141:542–555, 2016.
- [29] Simon Jenni and Hailin Jin. Time-equivariant contrastive video representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 9970–9980, October 2021.
- [30] Li Jing, Pascal Vincent, Yann LeCun, and Yuandong Tian. Understanding dimensional collapse in contrastive self-supervised learning. In International Conference on Learning Representations, 2022.
- [31] Hoel Kervadec, Jose Dolz, Meng Tang, Eric Granger, Yuri Boykov, and Ismail Ben Ayed. Constrained-cnn losses for weakly supervised segmentation. Medical image analysis, 54:88–99, 2019.
- [32] Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. Advances in Neural Information Processing Systems, 33:18661–18673, 2020.
- [33] Sun Hyung Kim, Vladimir S Fonov, Cheryl Dietrich, Clement Vachet, Heather C Hazlett, Rachel G Smith, Michael M Graves, Joseph Piven, John H Gilmore, Stephen R Dager, et al. Adaptive prior probability and spatial temporal intensity change estimation for segmentation of the one-year-old human brain. Journal of neuroscience methods, 212(1):43–55, 2013.
- [34] Pamela J LaMontagne, Tammie LS Benzinger, John C Morris, Sarah Keefe, Russ Hornbeck, Chengjie Xiong, Elizabeth Grant, Jason Hassenstab, Krista Moulder, Andrei G Vlassenko, et al. Oasis-3: longitudinal neuroimaging, clinical, and cognitive dataset for normal aging and alzheimer disease. MedRxiv, 2019.
- [35] Sangwoo Lee, Yejin Lee, Geongyu Lee, and Sangheum Hwang. Supervised contrastive embedding for medical image segmentation. IEEE Access, 9:138403–138414, 2021.
- [36] Bo Li, Wiro J Niessen, Stefan Klein, Marius de Groot, M Arfan Ikram, Meike W Vernooij, and Esther E Bron. Longitudinal diffusion mri analysis using segis-net: a single-step deep-learning framework for simultaneous segmentation and registration. NeuroImage, 235:118004, 2021.
- [37] Shijie Li, Neel Dey, Katharina Bermond, Leon Von Der Emde, Christine A Curcio, Thomas Ach, and Guido Gerig. Point-supervised segmentation of microscopy images and volumes via objectness regularization. In 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), pages 1558–1562. IEEE, 2021.
- [38] Shikun Liu, Shuaifeng Zhi, Edward Johns, and Andrew J Davison. Bootstrapping semantic segmentation with regional contrast. In International Conference on Learning Representations, 2022.
- [39] Jyrki MP Lötjönen, Robin Wolz, Juha R Koikkalainen, Lennart Thurfjell, Gunhild Waldemar, Hilkka Soininen, Daniel Rueckert, Alzheimer’s Disease Neuroimaging Initiative, et al. Fast and robust multi-atlas segmentation of brain magnetic resonance images. Neuroimage, 49(3):2352–2365, 2010.
- [40] Fausto Milletari, Nassir Navab, and Seyed-Ahmad Ahmadi. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In 2016 fourth international conference on 3D vision (3DV), pages 565–571. IEEE, 2016.
- [41] Andriy Myronenko. 3d mri brain tumor segmentation using autoencoder regularization. In International MICCAI Brainlesion Workshop, pages 311–320. Springer, 2018.
- [42] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
- [43] Jiahong Ouyang, Qingyu Zhao, Ehsan Adeli, Edith V Sullivan, Adolf Pfefferbaum, Greg Zaharchuk, and Kilian M Pohl. Self-supervised longitudinal neighbourhood embedding. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 80–89. Springer, 2021.
- [44] Taesung Park, Alexei A Efros, Richard Zhang, and Jun-Yan Zhu. Contrastive learning for unpaired image-to-image translation. In European Conference on Computer Vision, pages 319–345. Springer, 2020.
- [45] Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei A Efros. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2536–2544, 2016.
- [46] Oula Puonti, Juan Eugenio Iglesias, and Koen Van Leemput. Fast and sequence-adaptive whole-brain segmentation using parametric bayesian modeling. NeuroImage, 143:235–249, 2016.
- [47] Martin Reuter, Nicholas J Schmansky, H Diana Rosas, and Bruce Fischl. Within-subject template estimation for unbiased longitudinal image analysis. Neuroimage, 61(4):1402–1418, 2012.
- [48] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015.
- [49] Kendrick Shen, Robbie Jones, Ananya Kumar, Sang Michael Xie, Jeff Z HaoChen, Tengyu Ma, and Percy Liang. Connect, not collapse: Explaining contrastive learning for unsupervised domain adaptation. arXiv preprint arXiv:2204.00570, 2022.
- [50] Mark D Shen, Meghan R Swanson, Jason J Wolff, Jed T Elison, Jessica B Girault, Sun Hyung Kim, Rachel G Smith, Michael M Graves, Leigh Anne H Weisenfeld, Lisa Flake, et al. Subcortical brain development in autism and fragile x syndrome: evidence for dynamic, age-and disorder-specific trajectories in infancy. American Journal of Psychiatry, pages appi–ajp, 2022.
- [51] Feng Shi, Yong Fan, Songyuan Tang, John H Gilmore, Weili Lin, and Dinggang Shen. Neonatal brain image segmentation in longitudinal mri studies. Neuroimage, 49(1):391–400, 2010.
- [52] Feng Shi, Pew-Thian Yap, Guorong Wu, Hongjun Jia, John H Gilmore, Weili Lin, and Dinggang Shen. Infant brain atlases from neonates to 1-and 2-year-olds. PloS one, 6(4):e18746, 2011.
- [53] Yue Sun, Kun Gao, Zhengwang Wu, Guannan Li, Xiaopeng Zong, Zhihao Lei, Ying Wei, Jun Ma, Xiaoping Yang, Xue Feng, et al. Multi-site infant brain segmentation algorithms: The iseg-2019 challenge. IEEE Transactions on Medical Imaging, 40(5):1363–1376, 2021.
- [54] Meghan R Swanson, Mark D Shen, Jason J Wolff, Jed T Elison, Robert W Emerson, Martin A Styner, Heather C Hazlett, Kinh Truong, Linda R Watson, Sarah Paterson, et al. Subcortical brain and behavior phenotypes differentiate infants with autism versus language delay. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 2(8):664–672, 2017.
- [55] Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive multiview coding. In European conference on computer vision, pages 776–794. Springer, 2020.
- [56] Minh-Son To, Ian G Sarno, Chee Chong, Mark Jenkinson, and Gustavo Carneiro. Self-supervised lesion change detection and localisation in longitudinal multiple sclerosis brain imaging. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 670–680. Springer, 2021.
- [57] Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th international conference on Machine learning, pages 1096–1103, 2008.
- [58] Hongzhi Wang, Jung W Suh, Sandhitsu R Das, John B Pluta, Caryne Craige, and Paul A Yushkevich. Multi-atlas segmentation with joint label fusion. IEEE transactions on pattern analysis and machine intelligence, 35(3):611–623, 2012.
- [59] Jiangliu Wang, Jianbo Jiao, and Yun-Hui Liu. Self-supervised video representation learning by pace prediction. In European conference on computer vision, pages 504–521. Springer, 2020.
- [60] Wenguan Wang, Tianfei Zhou, Fisher Yu, Jifeng Dai, Ender Konukoglu, and Luc Van Gool. Exploring cross-image pixel contrast for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 7303–7313, October 2021.
- [61] Jie Wei, Feng Shi, Zhiming Cui, Yongsheng Pan, Yong Xia, and Dinggang Shen. Consistent segmentation of longitudinal brain mr images with spatio-temporal constrained networks. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 89–98. Springer, 2021.
- [62] Jure Zbontar, Li Jing, Ishan Misra, Yann LeCun, and Stephane Deny. Barlow twins: Self-supervised learning via redundancy reduction. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 12310–12320. PMLR, 18–24 Jul 2021.
- [63] Dewen Zeng, Yawen Wu, Xinrong Hu, Xiaowei Xu, Haiyun Yuan, Meiping Huang, Jian Zhuang, Jingtong Hu, and Yiyu Shi. Positional contrastive learning for volumetric medical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 221–230. Springer, 2021.
- [64] Guodong Zeng and Guoyan Zheng. Multi-stream 3d fcn with multi-scale deep supervision for multi-modality isointense infant brain mr image segmentation. In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pages 136–140. IEEE, 2018.
- [65] Richard Zhang, Phillip Isola, and Alexei A Efros. Split-brain autoencoders: Unsupervised learning by cross-channel prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1058–1067, 2017.
- [66] Qingyu Zhao, Ehsan Adeli, and Kilian M Pohl. Longitudinal correlation analysis for decoding multi-modal brain development. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 400–409. Springer, 2021.
- [67] Xiangyun Zhao, Raviteja Vemulapalli, Philip Andrew Mansfield, Boqing Gong, Bradley Green, Lior Shapira, and Ying Wu. Contrastive learning for label efficient semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 10623–10633, October 2021.
- [68] Yuanyi Zhong, Bodi Yuan, Hong Wu, Zhiqiang Yuan, Jian Peng, and Yu-Xiong Wang. Pixel contrastive-consistent semi-supervised semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 7273–7282, October 2021.
Checklist
-
1.
For all authors…
-
(a)
Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes] See abstract and Sec.1
-
(b)
Did you describe the limitations of your work? [Yes] See Sec. 5.
-
(c)
Did you discuss any potential negative societal impacts of your work? [Yes] See Sec. 5.
-
(d)
Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]
-
(a)
-
2.
If you are including theoretical results…
-
(a)
Did you state the full set of assumptions of all theoretical results? [N/A] No new theoretical results claimed.
-
(b)
Did you include complete proofs of all theoretical results? [N/A]
-
(a)
-
3.
If you ran experiments…
-
(a)
Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] Code: code with instructions is available at https://github.com/mengweiren/longitudinal-representation-learning. Data: public data links and procedures are described in the supplementary material.
- (b)
- (c)
-
(d)
Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See ‘Implementation details’ in Sec. 4.
-
(a)
-
4.
If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…
-
(a)
If your work uses existing assets, did you cite the creators? [Yes] See ‘Baselines and Evaluation Strategies’ in Sec. 4
-
(b)
Did you mention the license of the assets? [N/A]
-
(c)
Did you include any new assets either in the supplemental material or as a URL? [Yes] Our code is available freely at https://github.com/mengweiren/longitudinal-representation-learning.
- (d)
-
(e)
Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] See ‘Data and preprocessing’ in Sec. 4. Skull-stripping deidentifies human faces as only the brain is available.
-
(a)
-
5.
If you used crowdsourcing or conducted research with human subjects…
-
(a)
Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A]
-
(b)
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [Yes] See App. D.1.
-
(c)
Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A] See App. D.
-
(a)
Appendix A Additional Results and Experiments
Self-supervised similarity maps from anatomically-relevant key points are visualized in Figure 5. Using only self-supervision, our model learns semantically and positionally-aware representations.

Label-wise longitudinal consistency. To observe structure-specific performance, we report label-wise longitudinal consistency scores (as measured by STCS [36]) for OASIS3 and IBIS-subcort on the test set for the best performing models (as measured by Figure 4 of the main text) in Figure 6. IBIS-wmgm was not included as it only has two anatomical structures.

Few-shot and fully-supervised segmentation. While the main body of the paper presents results on one-shot segmentation (corresponding to the common real-world setting of single atlas-based segmentation), the proposed framework has benefits beyond one-shot segmentation. We present results obtained by using 10% and 100% of all the labeled training sets in Tables 3 and 4, respectively. As having more supervised data may necessitate changing our hyperparameters, we explore reducing the weight of the consistency-regularization from to for both experiments. In the 10% setting, we find that Ours w/ with a weight of obtains optimal performance and consistency. In the 100% setting, we find that including actually degrades performance and that optimal results are achieved with just the proposed pretraining (Ours w/o ). We hypothesize that these trends arise from imperfect deformable registration of the intra-subject images as they were warped considering only their intensity and not semantic structure (see App Section C.1). Therefore, having the imperfect self-supervision of in the low-annotation regime increases performance. However, hundreds of annotated samples remove the need for , while maintaining the segmentation performance benefits of our pretraining framework over existing work.
IBIS-subcort | |||||||
Model | Mean(std) Dice | Mdn Dice | Mean(std) IoU | Mdn HD95 | Mdn IoU | ASPC | STCS |
RandInitUnet.2D | 0.895(0.03) | 0.891 | 0.812(0.06) | 2.390 | 0.804 | 4.618 | 0.909 |
RandInitUnet.3D | 0.909(0.03) | 0.905 | 0.834(0.05) | 1.819 | 0.827 | 5.531 | 0.911 |
Context Restore [10] | 0.901(0.03) | 0.898 | 0.821(0.05) | 2.240 | 0.815 | 4.895 | 0.918 |
LNE [43] | 0.914(0.03) | 0.909 | 0.842(0.05) | 1.845 | 0.834 | 5.807 | 0.911 |
GLCL [9] | 0.905(0.03) | 0.900 | 0.827(0.05) | 2.486 | 0.819 | 4.369 | 0.913 |
PCL [63] | 0.904(0.03) | 0.900 | 0.826(0.05) | 2.881 | 0.819 | 4.479 | 0.911 |
PatchNCE [44] | 0.919(0.03) | 0.915 | 0.851(0.05) | 1.004 | 0.844 | 5.871 | 0.908 |
Ours w/o | 0.920(0.03) | 0.916 | 0.853(0.05) | 1.001 | 0.846 | 5.521 | 0.908 |
Ours w/ | 0.923(0.03) | 0.920 | 0.858(0.04) | 1.002 | 0.853 | 5.077 | 0.910 |
Ours w/ | 0.917(0.03) | 0.913 | 0.848(0.05) | 1.002 | 0.840 | 4.655 | 0.925 |
OASIS3 | |||||||
Model | Mean(std) Dice | Mdn Dice | Mean(std) IoU | Mdn HD95 | Mdn IoU | ASPC | STCS |
RandInitUnet.2D | 0.852(0.09) | 0.878 | 0.752(0.13) | 1.569 | 0.785 | 4.297 | 0.908 |
RandInitUnet.3D | 0.867(0.09) | 0.888 | 0.774(0.12) | 1.409 | 0.800 | 2.575 | 0.926 |
Context Restore [10] | 0.847(0.10) | 0.873 | 0.745(0.13) | 1.660 | 0.777 | 4.528 | 0.916 |
LNE [43] | 0.872(0.09) | 0.894 | 0.782(0.12) | 1.426 | 0.811 | 2.747 | 0.923 |
GLCL [9] | 0.861(0.09) | 0.884 | 0.765(0.12) | 1.387 | 0.794 | 4.005 | 0.911 |
PCL [63] | 0.862(0.09) | 0.885 | 0.766(0.12) | 1.328 | 0.796 | 3.896 | 0.910 |
PatchNCE [44] | 0.868(0.09) | 0.888 | 0.776(0.12) | 1.347 | 0.800 | 2.666 | 0.927 |
Ours w/o | 0.872(0.08) | 0.894 | 0.782(0.12) | 1.372 | 0.810 | 2.647 | 0.923 |
Ours w/ | 0.873(0.08) | 0.894 | 0.783(0.12) | 1.335 | 0.809 | 2.602 | 0.926 |
Ours w/ | 0.868(0.09) | 0.890 | 0.776(0.12) | 1.341 | 0.804 | 2.446 | 0.932 |
IBIS-subcort | |||||||
Model | Mean(std) Dice | Mdn Dice | Mean(std) IoU | Mdn HD95 | Mdn IoU | ASPC | STCS |
RandInitUnet.2D | 0.902(0.03) | 0.898 | 0.824(0.05) | 2.634 | 0.815 | 4.308 | 0.917 |
RandInitUnet.3D | 0.922(0.03) | 0.917 | 0.856(0.04) | 1.730 | 0.847 | 5.517 | 0.916 |
Context Restore [10] | 0.909(0.03) | 0.906 | 0.835(0.05) | 1.011 | 0.828 | 4.125 | 0.922 |
LNE [43] | 0.928(0.02) | 0.924 | 0.867(0.04) | 1.000 | 0.859 | 5.253 | 0.911 |
GLCL [9] | 0.914(0.03) | 0.909 | 0.843(0.05) | 1.011 | 0.834 | 4.191 | 0.917 |
PCL [63] | 0.913(0.03) | 0.909 | 0.842(0.05) | 1.006 | 0.833 | 4.122 | 0.918 |
PatchNCE [44] | 0.928(0.02) | 0.924 | 0.867(0.04) | 1.001 | 0.860 | 5.470 | 0.910 |
Ours w/o | 0.933(0.02) | 0.930 | 0.876(0.04) | 1.000 | 0.870 | 6.666 | 0.901 |
Ours w/ | 0.930(0.02) | 0.926 | 0.870(0.04) | 1.000 | 0.863 | 5.298 | 0.910 |
Ours w/ | 0.922(0.02) | 0.917 | 0.856(0.04) | 1.000 | 0.846 | 4.605 | 0.926 |
OASIS3 | |||||||
Model | Mean(std) Dice | Mdn Dice | Mean(std) IoU | Mdn HD95 | Mdn IoU | ASPC | STCS |
RandInitUnet.2D | 0.844(0.10) | 0.874 | 0.741(0.13) | 1.632 | 0.777 | 4.438 | 0.914 |
RandInitUnet.3D | 0.878(0.08) | 0.898 | 0.791(0.11) | 1.371 | 0.817 | 5.570 | 0.920 |
Context Restore [10] | 0.864(0.09) | 0.886 | 0.769(0.12) | 1.415 | 0.797 | 4.202 | 0.927 |
LNE [43] | 0.882(0.08) | 0.904 | 0.796(0.11) | 1.280 | 0.825 | 4.140 | 0.926 |
GLCL [9] | 0.864(0.09) | 0.887 | 0.769(0.12) | 1.424 | 0.799 | 3.590 | 0.926 |
PCL [63] | 0.865(0.09) | 0.890 | 0.771(0.12) | 1.315 | 0.802 | 3.328 | 0.926 |
PatchNCE [44] | 0.869(0.08) | 0.889 | 0.777(0.11) | 1.328 | 0.801 | 2.599 | 0.937 |
Ours w/o | 0.885(0.08) | 0.907 | 0.801(0.11) | 1.246 | 0.831 | 2.919 | 0.933 |
Ours w/ | 0.882(0.08) | 0.904 | 0.796(0.11) | 1.233 | 0.825 | 2.647 | 0.934 |
Ours w/ | 0.877(0.08) | 0.898 | 0.789(0.11) | 1.245 | 0.817 | 2.368 | 0.939 |
U-Net configuration. For consistency, we use the same base U-Net architecture for all baselines. Its configuration is modeled based on the one-shot segmentation mean Dice validation results on OASIS3 presented in Table 5 and evaluated over several training crop sizes, normalization layers, and channel width multipliers. Importantly, all configurations were trained without any pretraining to obtain a baseline. We observe that a random crop window of is optimal (rows A, B, C) and find that Instance Normalization (row D) instead of Batch Normalization degrades performance. Finally, we observe almost no performance degradation when reducing the model size via the channel width multiplier from 24 to 16 (row C vs E) and therefore use 16 for computational efficiency.
Exp | Crop size | Normalization | Channel width | mean(std) dice |
---|---|---|---|---|
A | batch norm | 24 | 0.768 ( 0.14) | |
B | batch norm | 24 | 0.795 ( 0.14) | |
C | batch norm | 24 | 0.786 ( 0.14) | |
D | instance norm | 24 | 0.777 ( 0.14) | |
E | batch norm | 16 | 0.792 ( 0.15) | |
F | batch norm | 8 | 0.786 ( 0.14) |
Training a randomly-initialized U-Net with . To investigate the standalone benefit of without any regularized pretraining, we apply to a randomly initialized 3D U-Net in the one-shot segmentation setting for IBIS-subcort and report its results in Fig. 7.

Using last decoder layer in loss. We investigate the impact of including the full-resolution feature (layer 21 in Table 9) in the EncDec loss configuration described in Section B.1. As shown in Figure 8, this notably degrades performance, indicating sensitivity to the exact layers used for calculation.

Additional modeling decisions. In addition to the quantitative ablations over our modeling choices included in the main paper and appendix, we make a few modeling decisions detailed in Table 6 based on qualitative assessments of the representation quality using feature similarity maps.
Parameter | Search space |
---|---|
# of layers in projector | 2,3 |
non-linearity | ELU, ReLU, LeakyReLU |
within subject registration | affine only, affine and deformable |
feature selection for similarity loss | conv output, activation output |
Additional one-shot quantification. As Table 1 and Figure 4 (main text) report median scores, Table 7 reports means and standard deviations for further interpretation. Similar to the medians, our method improves both mean performance and longitudinal consistency on IBIS-{wmgm, subcort}. On OASIS3, our Dice and IoU performance improves on all baselines except for LNE (with whom it is strongly competitive) and exceeds all baselines in terms of longitudinal consistency (ASPC, STCS).
IBIS-subcort | |||||
Model | Dice | IoU | HD95 | ASPC | STCS |
RandInitUnet.2D | 0.828(0.07) | 0.712(0.10) | 4.784(6.32) | 7.08(7.85) | 0.893(0.03) |
RandInitUnet.3D | 0.836(0.06) | 0.724(0.09) | 3.603(5.27) | 5.29(5.12) | 0.908(0.02) |
Context Restore [10] | 0.838(0.07) | 0.727(0.09) | 5.196(9.27) | 8.22(9.16) | 0.893(0.02) |
LNE [43] | 0.847(0.06) | 0.739(0.09) | 3.478(5.33) | 5.83(5.41) | 0.905(0.02) |
GLCL [9] | 0.835(0.07) | 0.722(0.10) | 3.640(4.16) | 5.47(6.23) | 0.905(0.03) |
PCL [63] | 0.833(0.07) | 0.720(0.09) | 3.528(3.56) | 5.78(6.10) | 0.902(0.03) |
PatchNCE [44] | 0.856(0.06) | 0.753(0.09) | 1.482(1.75) | 5.93(5.79) | 0.909(0.02) |
Ours w/o | 0.863(0.06) | 0.763(0.08) | 1.510(2.62) | 5.68(6.00) | 0.909(0.02) |
Ours w/ | 0.870(0.05) | 0.773(0.08) | 1.408(1.24) | 4.26(4.32) | 0.929(0.02) |
IBIS-wmgm | |||||
Model | Dice | IoU | HD95 | ASPC | STCS |
RandInitUnet.2D | 0.672(0.07) | 0.510(0.08) | 3.274(1.00) | 11.51(11.79) | 0.832(0.03) |
RandInitUnet.3D | 0.713(0.08) | 0.560(0.09) | 3.788(0.70) | 5.52(5.48) | 0.863(0.05) |
Context Restore [10] | 0.590(0.19) | 0.444(0.19) | 8.273(3.56) | 29.24(26.76) | 0.813(0.10) |
LNE [43] | 0.716(0.07) | 0.563(0.09) | 3.201(0.98) | 5.35(6.34) | 0.858(0.04) |
GLCL [9] | 0.707(0.05) | 0.550(0.06) | 4.112(1.14) | 8.47(8.85) | 0.858(0.02) |
PCL [63] | 0.715(0.08) | 0.562(0.09) | 4.974(1.75) | 10.65(13.29) | 0.868(0.05) |
PatchNCE [44] | 0.753(0.05) | 0.607(0.06) | 4.344(0.66) | 3.78(4.18) | 0.889(0.03) |
Ours w/o | 0.758(0.06) | 0.614(0.08) | 3.291(0.66) | 2.46(2.74) | 0.882(0.03) |
Ours w/ | 0.806(0.03) | 0.676(0.04) | 3.237(0.47) | 4.15(3.32) | 0.910(0.02) |
OASIS3 | |||||
Model | Dice | IoU | HD95* | ASPC | STCS |
RandInitUnet.2D | 0.781(0.15) | 0.661(0.17) | 3.425(5.00) | 10.27(28.84) | 0.873(0.05) |
RandInitUnet.3D | 0.804(0.13) | 0.689(0.16) | 3.424(5.85) | 3.53(6.35) | 0.923(0.05) |
Context Restore [10] | 0.801(0.14) | 0.687(0.16) | 4.841(10.00) | 6.88(15.21) | 0.899(0.04) |
LNE [43] | 0.813(0.13) | 0.700(0.15) | 3.171(5.50) | 8.81(31.94) | 0.898(0.04) |
GLCL [9] | 0.792(0.14) | 0.674(0.16) | 3.330(5.25) | 5.30(11.12) | 0.903(0.05) |
PCL [63] | 0.796(0.14) | 0.680(0.17) | 3.339(5.41) | 5.69(13.40) | 0.899(0.05) |
PatchNCE [44] | 0.809(0.13) | 0.697(0.16) | 3.298(5.93) | 3.49(6.70) | 0.920(0.04) |
Ours w/o | 0.813(0.13) | 0.702(0.15) | 3.289(5.68) | 3.40(6.11) | 0.923(0.04) |
Ours w/ | 0.810(0.13) | 0.697(0.16) | 3.733(6.41) | 2.76(4.96) | 0.934(0.03) |
Additional IBIS-wmgm segmentation results. For one-shot IBIS-wmgm segmentation, we train on a single 36 month old T1w/T2w MR image and quantitatively evaluate on 6 month old MR images (for additional motivation see Appendix Section D.2). In Figure 9, we visualize predictions using additional baselines and our framework on a held-out test subject. As the infant brain matures over time (Appendix Section D.2), the rows are arranged in order of the most to the least amount of biological domain-shift w.r.t. the 36 month-old training image. Further, as the 12, 24, and 36 month old held-out images do not have ground truth segmentations available, we are restricted to qualitative evaluations of segmentation quality.

Appendix B Additional Implementation Details
B.1 Additional Training Details

Figure 10 illustrates an in-depth overview of the proposed pretraining and representation learning framework. In the proposed loss configuration (EncDec), Pre-activation features from layers in Table 5 are sampled spatially to extract channel-wise vectors which are fed into the projector and predictor MLPs for calculation. Patch features from layer in the bottleneck (processed by the projector) are used for the orthogonality regularizer and layers in the decoder are used for the and regularizers. In the ablations in Table 2 (main text), the Enc loss setting refers to using features from layers for . Each iteration of pretraining loads three corresponding crops from two intra-subject images each for calculation. Higher batch sizes could have been used for the proposed loss functions but were not for consistency with one of our baselines (PatchNCE [44]) which is highly memory-intensive due to the need for a large number of negative samples alongside 3D computation.
During finetuning, the segmentation loss () is applied to the original unwarped images and the consistency loss () is applied to the nonlinearly aligned images. When calculating and , we perform individual forward/backward passes for each loss due to memory limitations as gradient accumulation over two forward passes was found to decrease performance on a validation set.
B.2 Architectures
The MLP architectures for the projector and predictor networks used for pretraining and representation learning are given in Table 8. The U-Net architecture used for both pretraining and finetuning is given in Table 9. During pretraining, an additional convolutional layer is used following layer 23 to reconstruct/denoise the input. During finetuning, a channel-wise softmax layer is attached following layer 23 for segmentation.
Projector (f) | ||
---|---|---|
id | Layer | Output size |
0 | FC(2048), BN, ReLU | (num patches batch size) 2048 |
1 | FC(2048), BN, ReLU | (num patches batch size) 2048 |
2 | FC(2048), BN | (num patches batch size) 2048 |
Predictor (p) | ||
id | Layer | Output size |
0 | FC(256), BN, ReLU | (num patches batch size) 256 |
1 | FC(2048), BN, ReLU | (num patches batch size) 2048 |
id | Layer | Output size |
---|---|---|
0 | Conv3D(nc), BN, ReLU | bs, w, h, d, nc |
1 | Conv3D(nc), BN, ReLU | bs, w, h, d, nc |
2 | Conv3D(nc), BN, ReLU | bs, w, h, d, nc |
3 | MaxPool(2), Conv3D(2nc), BN, ReLU | bs, w/2, h/2, d/2, 2nc |
4 | Conv3D(2nc), BN, ReLU | bs, w/2, h/2, d/2, 2nc |
5 | MaxPool(2), Conv3D(4nc), BN, ReLU | bs, w/4, h/4, d/4, 4nc |
6 | Conv3D(4nc), BN, ReLU | bs, w/4, h/4, d/4, 4nc |
7 | MaxPool(2), Conv3D(8nc), BN, ReLU | bs, w/8, h/8, d/8, 8nc |
8 | Conv3D(8nc), BN, ReLU | bs, w/8, h/8, d/8, 8nc |
9 | MaxPool(2), Conv3D(16nc), BN, ReLU | bs, w/16, h/16, d/16, 16nc |
10 | Conv3D(16nc), BN, ReLU | bs, w/16, h/16, d/16, 16nc |
11 | Upsample(2), Concatenate with 8 | bs, w/8, h/8, d/8, 24nc |
12 | Conv3D(16nc), BN, ReLU | bs, w/8, h/8, d/8, 8nc |
13 | Conv3D(16nc), BN, ReLU | bs, w/8, h/8, d/8, 8nc |
14 | Upsample(2), Concatenate with 6 | bs, w/4, h/4, d/4, 12nc |
15 | Conv3D(4nc), BN, ReLU | bs, w/4, h/4, d/4, 4nc |
16 | Conv3D(4nc), BN, ReLU | bs, w/4, h/4, d/4, 4nc |
17 | Upsample(2), Concatenate with 4 | bs, w/2, h/2, d/2, 6nc |
18 | Conv3D(2nc), BN, ReLU | bs, w/2, h/2, d/2, 2nc |
19 | Conv3D(2nc), BN, ReLU | bs, w/2, h/2, d/2, 2nc |
20 | Upsample(2), Concatenate with 2 | bs, w, h, d, 3nc |
21 | Conv3D(nc), BN, ReLU | bs, w, h, d, nc |
22 | Conv3D(nc), BN, ReLU | bs, w, h, d, nc |
23 | Conv3D(n) | bs, w, h, d, n |
B.3 Baseline Reimplementation Details
LNE. LNE [43] is a self-supervised global representation learning method that models longitudinal effects in an autoencoder latent space and is repurposed here for longitudinal segmentation. During pretraining, we require the same base U-Net architecture for consistency with other baselines. Therefore, to extract a 1024D global representation as in the original paper, we add a global pooling layer on the output of layer in Table 9, followed by a single 1024D fully-connected layer. With respect to their hyperparameters, due to the larger base network size, we reduce the batch size from 64 to 16 images due to memory considerations. We maintain the other hyperparameters of LNE including downsampling the images to , using neighborhoods, and weighing auxiliary losses as and (see original work for definitions [43]).
PatchNCE. PatchNCE [44] proposes a framework for unsupervised multi-scale patchwise contrastive learning applied to the problem of unpaired image translation and is extended to serve as a baseline in this work as it shares a partially similar motivation. To adapt it to our generic longitudinal representation learning setting, we impose the same modeling assumption as our method (spatial indices in correspondence across an image time-series are positive samples and all other spatial indices are negatives) and modify both its data sampling and loss calculation. First, we perform forward passes on corresponding intra-subject crops. At each selected layer, feature vectors are randomly sampled for the longitudinal InfoNCE [32] loss: , where for any query feature index , is the set of indices of all positives (from the same subject at the same spatial index) and is the set of all query indices (with size , where is the number of timepoints in the batch and is the number of patches). We choose timepoints and patches in our experiments due to memory limits. An ablation study of PatchNCE over the layers used for the loss function and the temperature is given in Table 11. For prototyping efficiency, we perform baseline tuning using a three-layer MLP of size 256 and use validation Dice as the model selection criterion to choose as our final configuration. For fair comparison with our model, we increase the MLP width in the PathNCE model to 2048 in all other experiments reported, although this does not significantly alter results.
Context Restoration. [10] proposes context restoration as a pretext task for self-supervised representation learning which consists of randomly and repeatedly swapping image patches and training the network to restore the original image. We extend Algorithm 1 in [10] to use 3D inputs and targets and tune it with a varying number of swapping iterations (#swaps) and swapped patch sizes. Downstream validation dice scores are provided in Table 11 and configuration is chosen as the final model.
PCL, GLCL, and 2D U-Net. PCL and GLCL [9, 63] are 2D slice-based contrastive learning methods designed for 3D volume segmentation. We follow the optimal configurations (e.g. temperature, number of partitions, etc.) reported in the original papers. As these methods use a 2D U-Net trained on 2D slices, we additionally benchmark against a randomly initialized 2D U-Net to obtain a baseline for 2D methods. We use a batch size of 128 for all 2D baselines. Within each batch, we compare two different data sampling strategies: (1) randomly sampling 2D slices across all 3D volumes; (2) sampling intra-volume slices within each batch (i.e. treat one of the dimensions from a crop as the batch dimension for 2D networks). We train a randomly initialized 2D U-Net for the one-shot segmentation task, and found that (2) significantly increases the validation mean Dice over (1) on OASIS3 from to under the same architecture and optimization set up. We therefore use batch sampling strategy (2) for all 2D model benchmarks (PCL, GLCL, RandInitUnet.2D).
Exp | layers | temperature | Mean(std) dice |
---|---|---|---|
A | Enc | 0.07 | 0.834 ( 0.060) |
B | Enc | 0.1 | 0.834 ( 0.062) |
C | Enc | 0.2 | 0.823 ( 0.067) |
D | Enc | 0.5 | 0.830 ( 0.064) |
E | EncDec | 0.07 | 0.854 ( 0.056) |
F | EncDec | 0.1 | 0.856 ( 0.054) |
G | EncDec | 0.2 | 0.846 ( 0.058) |
H | EncDec | 0.5 | 0.850 ( 0.054) |
Exp | patch size | # swaps | Mean(std) Dice |
---|---|---|---|
A | 10 | 0.778 ( 0.141) | |
B | 10 | 0.779 ( 0.142) | |
C | 30 | 0.775 ( 0.138) | |
D | 50 | 0.782 ( 0.135) | |
E | 100 | 0.778 ( 0.160) |
Appendix C Additional data preparation details
C.1 Registration
C.1.1 Affine registration
To warp all images to a common space, we warp all images affinely to a constructed template [4] (with an affine transformation model) using ANTs
111https://github.com/ANTsX/ANTs with the following command:
antsMultivariateTemplateConstruction2.sh \ -d 3 \ -o OUTPUT_FOLDER/T \ -i 1 -g 0.2 -j 128 -c 2 -r 1 -n 0 -m MI -l 1 \ -t Affine INPUT_FOLDER/*t1.nii.gz’
Once affinely aligned on a dataset-wide level, we proceed with longitudinal intra-subject deformable alignment for the calculation of and , as described below.
C.1.2 Nonlinear Deformable Registration
Proper spatial correspondence of positive samples for and label maps for (the similarity and consistency losses, respectively) requires nonlinear/deformable registration of all intra-subject images to a common reference. To this end, we employ the ANTs
SyN algorithm from [3] to register within-subject images to a single timepoint in a series of acquisitions with the following command:
fix = INPUT_FOLDER/{subj}_{trg_tp}_t1.nii.gz for src_tp in tps: moving = INPUT_FOLDER/{subj}_{src_tp}_t1.nii.gz antsRegistration \ --verbose 1 \ --dimensionality 3 \ --float 1 \ --output [OUTPUT_FOLDER/{subj}_{src_tp}_to_{trg_tp}_t1_, \ OUTPUT_FOLDER/{subj}_{src_tp}_to_{trg_tp}_t1_Warped.nii.gz, \ OUTPUT_FOLDER/{subj}_{src_tp}_to_{trg_tp}_t1_InvWarped.nii.gz] \ --transform SyN[0.15, 9, 0.2] \ --metric CC[{fix}, {moving}, 1, 2, Random, 0.4] \ --convergence [250x125x50, 1e-5, 10] \ --shrink-factors 4x2x1 \ --smoothing-sigmas 2x1x0vox \ --interpolation Linear
where fix is the intra-subject image from a selected timepoint trg_tp to which all other timepoints src_tp are registered.
C.2 Additional preprocessing and label generation
With respect to OASIS3, the publicly-available dataset arrives preprocessed with cross-sectional FreeSurfer processing (intensity and geometric normalization and segmentation). As cross-sectional FreeSurfer does not account for longitudinal effects [47], FreeSurfer V6.0 is used for additional longitudinal processing on top of the publicly-released cross-sectional FreeSurfer segmentations. A comparison between longitudinal and cross-sectional FreeSurfer analysis based on training set is given in Fig. 11, where we observe notable improvements to longitudinal segmentation consistency (as quantified by STCS [36]) over all structures.

For IBIS-{subcort, wmgm}, all T1w/T2w MR images are preprocessed following standard procedures described in [50] of gradient distortion correction, bias-field correction, within-subject and within time-point multi-modality registration, and brain extraction. For IBIS-subcort, ground truth label generation follows the segmentation procedures outlined in [50]. For IBIS-wmgm, we use segmentations generated by a fully-supervised network trained on an external dataset, see Appendix Section D.2 for more details.
Once preprocessed and aligned with methods described in Section C.1, the images are cropped to a common field-of-view ( for IBIS, for OASIS3, corresponding to smaller brain volumes for infants vs. adults) for compatibility with common multi-resolution neural network architectures.
C.3 Data splitting with repeat acquisitions

IBIS | ||||
---|---|---|---|---|
IBIS-subcort | IBIS-wmgm | |||
Split | (N_cross, M_cross) | (N_long, M_long) | (N_cross, M_cross) | (N_long, M_long) |
Total | (552, 1272) | (455, 1175) | (552, 1272) | (552, 1272) |
Train | (386, 887) | (313, 814) | (386, 887) | (313, 814) |
Validation | (55, 133) | (50, 128) | (55, 42) | (50, 128) |
Test | (111, 252) | (92, 233) | (111, 80) | (92, 233) |
OASIS3 | ||
---|---|---|
Split | (N_cross, M_cross) | (N_long, M_long) |
Total | (992, 1639) | (422, 1069) |
Train | (694, 1147) | (293, 746) |
Validation | (98, 166) | (48, 116) |
Test | (200, 326) | (81, 207) |
We perform a 70-10-20 train, validation, and test split on subject-wise basis from a total of subjects and images ( due to image time-series acquisitions), with the entire procedure illustrated in Figure 12. This level of the data split hierarchy ( subjects and images) is used for segmentation training and evaluation. For longitudinal pretraining and longitudinal consistency evaluation (ASPC, STCS in Table 1 and Figure 4 of the main text), we further filter subjects with at least 2 acquisitions per subject to obtain images.
Finally, IBIS-subcort and IBIS-wmgm are segmentation tasks which share the same unlabeled pretraining data. Importantly, we note that as IBIS-wmgm is evaluated and tested only on 6-month-old MR images (representing the strongest domain shift, see Appendix Section D.2 for its motivation), its segmentation validation and test sets contain fewer images than IBIS-subcort. Table 12 provides the exact sample sizes for each task.
Appendix D Miscellaneous Details
D.1 Data availability and IRB information
IBIS. The IBIS/Autism MRI data is available through NIH NDA222https://nda.nih.gov/edit_collection.html?id=19. The Infant Brain Imaging Study (IBIS) Network is a National Institutes of Health–funded Autism Center of Excellence project and consists of a consortium of eight universities in the United States and Canada. Parents provided informed consent and the institutional review board at each site approved the research protocol.
OASIS-3. OASIS-3 [34] is publicly available through the OASIS webpage333https://www.oasis-brains.org/. Ethical approval was obtained by the relevant ethics committees and informed consent was obtained from all participants following procedures set by the IRB at the Washington University School of Medicine.
D.2 IBIS-wmgm segmentation task details

Infant white matter maturation. Neurodevelopment in infants and toddlers is a highly-complex process of macro and micro-structural changes (for example, growth and myelination, respectively). Relevant to the scope of this paper, real-world tissue segmentation (into grey and white matter) of infant brains at approximately 6 months of age (post-birth) is highly difficult due to ongoing white matter myelination giving the brain an “isointense" (flat in intensity) appearance which ambiguates anatomical edges. Fig. 13 visualizes this phenomenon.
Currently, existing datasets for supervised segmentation of isointense infant brains [53] construct their ground truth labels by first nonlinearly registering algorithmic segmentations of an older timepoint from a given subject image time-series (which is straightforward to segment with existing tools) and then manually correcting the warped labels with the expertise of neuroradiologists.
In our one-shot segmentation setting, we take a partially analogous approach to segmentation of isointense infant brains. We first algorithmically segment [46] a single arbitrarily-selected 36 month old T1w/T2w MR image and use this reliable segmentation to train all segmentation baselines and the proposed method. These methods which have only been trained on a 36 month old brain are then quantitatively evaluated on isointense 6 month old brain tissue segmentation in this paper. The target labels for the 6 month old evaluation set are generated algorithmically with supervised segmentation networks following [64] which have been trained on a separate dataset [53].
Appendix E Need for regularization
In this section, we qualitatively and quantitatively illustrate that proper regularization of patch-wise negative-free multiscale representation learning avoids low-diversity or degenerate decoder representations which hamper downstream segmentation performance. We primarily compare the complete proposed model (‘Ours w/ regularization’) against an ablation (‘Our ablation w/o regularization’) that only uses the patch-wise similarity loss and does not use the denoising, variance, covariance, and orthogonality regularization (which corresponds to ablation E from Table 2).
Figure 14 visualizes activations from layers {6,8,10,13,16} of Table 9 (left) and the sorted singular values of the spatial covariance matrices of multi-layer feature projections from the corresponding layer-wise projector output (right) trained with/without regularization. Given a flattened mid-axial feature projection (where and is the vectorized spatial dimensionality), we calculate the spatial covariance as , where is a spatial feature vector and . The lower the rank of , the lower the spatial variability of representations.




