Supplementary Material: FSL Framework to Reduce Inter-Observer Variability
I The ParESN Model
The Parallel ESN Framework presented in this work is inspired by the previous works in [sohiniesn] and [esn2]. The primary difference with the model in [sohiniesn] is the variation in importance for the hidden states with respect to a previous pixel in the same image vs. the same pixel in the previous image. With respect to [esn2], the parallel branches in this work generate similarly trained regional proposals (RPs) instead of a cumulative combination from the three parallel streams in [esn2]. Also, the choice of three parallel layers in this work is optimal amongst the choice of parallel RPs based on a grid search and leave-one-out cross validation across vendor image stacks. The detailed system setup is shown in Fig. 1.

The OCT cyst images per vendor stack represent a volumetric level scan. This implies if a large cyst appears in a scan, there is a high probability of the same cyst appearing in some shape and form in the previous and succeeding scans as well. The proposed setup is analogous to the video processing setup in [sohiniesn]. Thus, the reservoir states per pixel location of subsequent images will be affected by the previous and current image, represented in (1).
(1) | |||
At the end of the training stage, are computed for each parallel layer using (2).
(2) |
where, are the extended system states for the 3 parallel layers, evaluated over training images, and represents the target label at pixel location , and represents identity matrix.
The leave-one-out cross validation experiment across vendor stacks helps identify the optimal parameter set in (1), (2), respectively. We observe to be optimal from the search set of in increments of 0.02. This implies very low “leaky-memory” requirement for the data set [DeepESN]. Also, the sensitivity to is found to be very low, with being optimal for the setup in the search set of in order increments of 10.
II Qualitative Analysis of Target Label Selection Algotithm (TLSA)
The key contribution of our work is the use of the RPs to detect the “best” manual annotation at image level. Examples of the TLSA for TL vs noisy TL selection and for vs. selection are shown below.
II-A Examples of TLSA against noisy labels
For this experiment the Random Crop and Paste function is invoked to generate noisy TLs. In Fig 2, the actual TLs are represented as red, noisy generations using RCAP as blue and their intersections as white regions, respectively. Also, the RPs are represented in red, green and blue planes, respectively. In Fig. 2 qualitative explanations are provided for manual interventions and for automatic selection of actual TL over the noisy counterparts.


II-B Examples of TLSA for best TL selection
In Fig. 3, qualitative examples of best TL selection are shown. In Fig. 3, the left columns represent the TLs such that is in red, in blue and their intersection is in white. The right columns represent the RPs in red, green and blue planes respectively. In Fig. 3(a), the Few-shot learning (FSL) models are trained on , but the RPs agree more with , and hence, is selected as the best label for all these images. In Fig. 3(b), the FSL models are trained on , but the RPs agree more with , and hence is selected as the best label for all these images. The examples in Fig. 3 demonstrate the extent of variabilities by manual target labels and .
For the images in Fig 3(a), annotations represented many contiguous small cyst regions, based on which, RPs get trained to identify small contiguous cysts in test images. However, annotates some larger cyst areas in the test images as opposed to that detects smaller cysts. Thus, is preferable in all such examples in Fig 3(a). For images in Fig. 3(b), the RPs trained on have an affinity to detect large cysts that appear in . Hence, in all these examples, is the preferred TL in spite of the FSL models being trained on . From the examples in Fig. 3, we are able to qualitatively assess the importance of the TLSA for standardizing cyst segmentation to overcome inter-observer variabilities.

