This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Light-weight Retinal Layer Segmentation with Global Reasoning

Xiang He, Weiye Song, Yiming Wang, Fabio Poiesi, Ji Yi, Manishi Desai, Quanqing Xu, Kongzheng Yang, and Yi Wan This work was supported in part by the National Natural Science Foundation of China under Grant 51975336, and 62205181, in part by the Natural Science Foundation of Shandong Province under Grant ZR2022QF017, in part by the Natural Science Outstanding Youth Fund of Shandong Province under Grant 2023HWYQ-023, in part by the Taishan Scholar Foundation of Shandong Province under Grant tsqn202211038, in part by the NIH under Grant R01NS108464, in part by the Key Technology Research and Development Program of Shandong Province under Grant 2020JMRH0202 and 2022CXGC020701, in part by the Shandong Province New Old Energy Conversion Major Industrial Tackling Projects under Grant 2021-13, in part by the Key Research and Development Project of Jining City under Grant 2021DZP005, and in part by the Shandong University Education Teaching Reform Research Project under Grant 2022Y133, 2022Y124, and 2022Y312. (Corresponding author: Yi Wan.)Xiang He is with the School of Mechanical Engineering, and also with the Joint SDU-NTU Centre for Artificial Intelligence Research (C-FAIR), Shandong University, Jinan, Shandong, China (e-mail: [email protected]).Weiye Song, Quanqing Xu, Kongzheng Yang, and Yi Wan are with the School of Mechanical Engineering, Shandong University, Jinan 250061, Shandong, China (e-mail: [email protected]; [email protected]; [email protected]; [email protected]).Yiming Wang, Fabio Poiesi are with the Fondazione Bruno Kessler, Via Sommarive 18, Povo, TN 38123, Italy (e-mail: [email protected]; [email protected]).Ji Yi is with the Department of Biomedical Engineering and the Department of Ophthalmology, Johns Hopkins University, Baltimore, Maryland 21231, USA (e-mail: [email protected]).Manishi Desai is with the Department of Ophthalmology, Boston University School of Medicine, Boston Medical Center, Boston 02118, USA (e-mail: [email protected]).
Abstract

Automatic retinal layer segmentation with medical images, such as optical coherence tomography (OCT) images, serves as an important tool for diagnosing ophthalmic diseases. However, it is challenging to achieve accurate segmentation due to low contrast and blood flow noises presented in the images. In addition, the algorithm should be light-weight to be deployed for practical clinical applications. Therefore, it is desired to design a light-weight network with high performance for retinal layer segmentation. In this paper, we propose LightReSeg for retinal layer segmentation which can be applied to OCT images. Specifically, our approach follows an encoder-decoder structure, where the encoder part employs multi-scale feature extraction and a Transformer block for fully exploiting the semantic information of feature maps at all scales and making the features have better global reasoning capabilities, while the decoder part, we design a multi-scale asymmetric attention (MAA) module for preserving the semantic information at each encoder scale. The experiments show that our approach achieves a better segmentation performance compared to the current state-of-the-art method TransUnet with 105.7M parameters on both our collected dataset and two other public datasets, with only 3.3M parameters.

Index Terms:
retinal layer segmentation, light-weight, multi-scale asymmetric attention, visible-light OCT.

I INTRODUCTION

With the accelerated pace of lifestyle and the popularity of electronic products, many people lack awareness of eye protection, leading to an increasing incidence of retinal diseases. Many retinal diseases are accompanied by changes in the thickness of the retinal layers. For instance, in patients with glaucoma, the retinal nerve fiber layer is thinner and the optic disc is atrophied compared to the healthy eyes [1, 2]. In diabetic retinopathy, the macular region of the retina thickens and becomes edematous [3]. Current research indicates that the changes in the retinal layer begin to occur in the early stages of the disease[4, 5], therefore, automatic analysis of the retinal layers and monitoring their morphological changes can help to understand the disease progression and provide early treatment.

OCT can be used for non-invasive 3D structural imaging of the retina [6], with which experienced ophthalmologists can manually inspect the retinal layers to identify the disease progression. However, manual inspection is inefficient and tedious. It is also challenging for various levels of ophthalmologists to ensure consistency and objectivity of the inspection. Although imaging techniques with a higher resolution are available now, such as Visible-light OCT images [7, 8, 9], this does not address completely the above status quo. Recent convolutional neural networks (CNNs), as typified by U-net [10], have achieved promising results regarding semantic segmentation in the field of medical image analysis. The seminar work ReLayNet [11] has triggered widespread interest in applying CNNs to segment retinal layers. ReLayNet follows a multi-scale encoder-decoder structure, which has inspired many mainstream semantic segmentation frameworks to adopt, such as U-net [10] and Attention_Unet [12]. However, such structure can be limited in two main aspects. Firstly, most U-shape encoder-decoder structures perform the multi-scale feature extraction followed by feature fusion via only residual connections. Thus, it is limited to maintain and fully exploit the semantic information of feature maps at all scales, resulting in false positives in the background region when performing retinal layer segmentation. Secondly, many of these methods do not pay attention to the actual application needs of OCT devices, often pursuing accuracy unilaterally, for example, such multi-scale reasoning often comes at the cost of a large number of parameters, which can be computationally demanding for the deployment of real-time clinical applications.

To overcome the above-mentioned limitations, we propose LightReSeg, a novel multi-scale encoder-decoder network for end-to-end retinal layer segmentation, which is a light-weight segmentation approach tailored to practical OCT device requirements. In order to reduce segmentation errors in the background region, we propose to incorporate a Transformer block [13] to features with global reasoning and to introduce a multi-scale asymmetric attention (MAA) module to better preserve the semantic information at each encoder scale. To achieve computation efficiency, both the backbone and the MAA module adopt light-weight feature extractors, such as depthwise separable convolution and asymmetric convolution [14, 15]. LightReSeg achieves the state-of-the-art segmentation accuracy with only 3.3M parameters, a significantly light model, as shown in Fig. 1.

Refer to caption
Figure 1: The model size v.s. the segmentation accuracy in terms of mIoU of the state-of-the-art retinal layer segmentation methods. Our LightReSeg achieves the highest mIoU compared to SOTA methods while maintaining a smaller model size.

In summary, our contributions are:

  • We propose LightReSeg, a novel encoder-decoder structure for retinal layer segmentation, which exploits global information in a light-weight manner, to improve the segmentation accuracy with reduced computational complexity, it provides valuable experience for algorithms designed for practical applications in medical devices.

  • We propose a novel attention module, the MAA module, to jointly work with Transformer, in order to address erroneous segmentation in the background area by allowing the model to reason in a global manner.

  • We perform our method on the visible-light OCT images for the first time and score the best segmentation performance, it provides experience for the algorithm to perform robustly on datasets from different domains.

This paper is organized as follows: Sec. II describes the research progress in the field of biological tissue segmentation and retinal layer segmentation. Sec. III introduces the overall framework, the feature extractor, MAA module and the light-weight designs of the approach. Sec. IV describes the datasets, performance metrics, implementation details, analyzes quantitatively and qualitatively the experimental results, and performs ablation experiments. Finally, the conclusions of the paper are given in Sec. V.

II RELATED WORK

In this section, we first provide an overview of the research progress of semantic segmentation, especially light-weight semantic segmentation methods, followed by a more specific coverage of deep-learning-based retinal layer segmentation.

II-A Medical Image Segmentation

Medical image segmentation has made rapid progress under the promotion of deep learning, especially in the fields of biological tissue recognition and lesion detection, which has sparked a large amount of research on medical image segmentation. The morphology of the optic disc and optic cup, as well as the cup-to-disc ratio, can be used to assess glaucoma, Wang et al. proposed an automatic segmentation approach based on CNNs to accurately segment the optic disc and optic cup from fundus images for glaucoma detection[16]. Precise segmentation of retinal vessels from fundus images is essential for intervention in numerous diseases and high-precision segmentation of retinal vessels still remains a challenging task, Li et al. proposed a dual-path progressive fusion network, named DPF-Net, which achieved better segmentation ability by effectively fusing global and local features[17]. Abdominal multi-organ image segmentation plays a crucial role in the diagnosis and treatment of many diseases. traditional methods of manual depiction are inefficient and subjective, therefore, Chen et al. proposed TransUnet based on CNN and Transformer block for abdominal multi-organ segmentation[18]. TransUnet was the first model to introduce the Transformer block into the U-shaped encoder-decoder structure and achieved very high accuracy in segmentation. There are also segmentation studies dedicated to lesion regions, such as Astaraki et al. who segmented different types of lung pathological regions to enable screening for lung diseases[19]. In addition, some lightweight segmentation methods for practical applications have also been proposed, such as SegFormer[20], A-net[21], EMV-Net[22].

Refer to caption
Figure 2: The network of LightReSeg follows a U-shape encoder-decoder structure. The encoder takes as the input a retinal image of dimension (3,H,W)(3,H,W) and performs multi-scale feature extraction that outputs feature maps of NN scales, where NN is set to 4 in our design. The last feature map is further fed to the Transformer Layers through a linear transformation to extract features that reasons in long-range to help the reduction of the segmentation errors in the background region. The resulted features after the Transformer Layers are then fused via reshaping and up-sampling with the multi-scale encoder features that are optimized by the proposed MAA module. The final fused feature map FoutF_{out} is exploited for the retinal layer segmentation via convolutions.

II-B Retinal Layer Segmentation

The morphology of the retinal layer is associated with the development of many ophthalmic diseases, therefore, there are many studies on retinal layer segmentation based on CNN and OCT images. The success of ReLayNet opened the door to the application of CNNs for retinal layer segmentation[11], a typical U-shaped network based on an encoder-decoder structure, which achieved the best performance for retinal layer segmentation at that time. Since layer segmentation of OCT retinal images is prone to speckle noise, intensity inhomogeneity, Wang et al. proposed a boundary-aware U-Net for retinal layer segmentation by detecting accurate boundaries[23]. To solve the topological errors caused by not considering the order of retinal layers, Liu et al proposed a novel deep learning-based framework that employed the distance maps of layer surfaces to convert the layer segmentation task into a multitasking problem for classification and regression[24]. To ensure the continuity of the retinal layer boundary and obtain more accurate segmentation results, Hu et al proposed a coarse-to-fine retinal layer boundary segmentation method based on the embedded residual recurrent network and the graph search, it has achieved good performance in both qualitative and quantitative indicators[25]. In addition, considering the impact of different neurological diseases on the retinal layer, Gende et al presented a fully automatic approach for the retinal layer segmentation in multiple neurodegenerative disorder scenarios[26], the results indicate that the model is more robust. There are also studies aimed at practical application deployment, such as He et al. proposed a light-weight retinal layer segmentation approach that achieved good performance with a lower number of parameters[22].

III METHODOLOGY

III-A Framework Overview

Our proposed LightReSeg is a U-shape network based on an encoder-decoder structure as shown in Fig. 2. The network takes a retinal layer OCT image as input. The first NN scale feature maps (F1F_{1}, …, FN1F_{N-1}, FNF_{N}) are extracted by the encoder which consists of a multi-scale feature extractor, respectively. In order to further deepen the global reasoning capability in the depth direction, the feature map FNF_{N}, which has to pass through a depth feature encoder: Transformer block[13]. The feature map FTNF_{TN} further encoded by the Transformer block has no change in shape and size, but it facilitates long-range global reasoning. To better preserve details at different scales, multi-scale features (F1F_{1}, …, FN1F_{N-1}, FNF_{N}) are then merged with its corresponding layers in the decoder part via our proposed MAA module, which outputs the feature maps (FM1F_{M1}, …, FMN1F_{MN-1}, FMNF_{MN}) with the same shape size as its input. The FTNF_{TN} output from the Transformer block is merged with the FMNF_{MN} for feature fusion, and then after 2 fold size up-sampling, it continues to fuse with FMN1F_{MN-1} for feature fusion and then follows a similar operation until the fusion of the up-sampled restored feature map with FoutF_{out} is completed, and after channel adjustment, the final retinal layer segmentation image is output. LightReSeg is designed for real-time clinical operations, where light-weight designs are employed where possible.

We present in full the multi-scale encoder in Section III-B followed by the proposed MAA module that serves for multi-scale feature fusion in Section III-C. Finally, we describe the design details for computation efficiency in Section III-D.

III-B Multi-scale Encoder

The multi-scale encoder extracts NN scale features to fully exploit the semantic information of feature maps at all scales, further, in order to make the features have better global reasoning capabilities, we incorporate a Transformer block[13], which is a module that further optimizes the NNth scale feature map extracted from the deepest scale of the encoder. To ensure the dimensionality is consistent with the Transformer block port, we pre-process the feature map FNF_{N} before being processed by the Transformer block. The feature map FNF_{N} is first converted into a 2D2D patch sequence as

FNFN={xpkRP2×c|k=1,,Z},Z=HWP2,F_{N}\Rightarrow F_{N}^{{}^{\prime}}=\left\{\right.x_{p}^{k}\in R^{P^{2}\times c}~{}|~{}k=1,...,Z\left.\right\}~{},~{}Z=\frac{HW}{P^{2}}, (1)

where cc is the number of channels in the FNF_{N} feature map, PP is the size of the patch, and HH and WW are the height and width of the FNF_{N} feature map.

We use a trainable linear projection to map the vector patch xpx_{p} to a latent D-dimensional embedding space. To encode the spatial information of the patches, we learn specific position embeddings that are added to preserve the positional information in the patch embedding as

z0=[x0;xp1E;xp2E;;xpNE]+Epos,z_{0}=\left[x_{{0}};x_{p}^{1}E;x_{p}^{2}E;\cdots;x_{p}^{N}E\right]+E_{pos}, (2)

where E(P2C)×DE\in\mathbb{R}^{\left(P^{2}\cdot C\right)\times D} represents the patch embedding projection, the x0x_{0} is an additional learnable vector concatenated with the remaining vectors to integrate the information of all remaining vectors, the [;;]\left[\cdot;\cdots;\cdot\right] is the concatenation operator, and Epos(Z+1)×DE_{pos}\in\mathbb{R}^{(Z+1)\times D} represents the position embedding[13].

Refer to caption
Figure 3: The structure of Multi-scale Asymmetric Attention module.

Then, inputting z0z_{0} into the Transformer Layer, as shown in Fig. 2, first goes through the first Layer Norm (LN) layer and then enters the Multi-head Self-attention (MSA) layer, here MSA(LN(z0))+z0MSA\left(LN\left(z_{0}\right)\right)+z_{0} forms the residual structure and get the z0z_{0}^{\prime}. Then, after passing through the second LN layer, it enters the Multi-Layer Perceptron (MLP) layer, here MSA(LN(z0))+z0MSA\left(LN\left(z_{0}^{\prime}\right)\right)+z_{0}^{\prime} once again forms the residual structure and get the z1z_{1}.

At this point, the first Transformer layer is finished. We have set three Transformer layers in our approach, so we have to cycle three times.

zL=[x0;xp1;xp2;;xpN][xp1;xp2;;xpN]\begin{split}z_{L}=\left[x_{{0}}^{{}^{\prime}};x_{p}^{1^{\prime}};x_{p}^{2^{\prime}};\cdots;x_{p}^{N^{\prime}}\right]\Rightarrow\left[x_{p}^{1^{\prime}};x_{p}^{2^{\prime}};\cdots;x_{p}^{N^{\prime}}\right]\end{split} (3)

When the zLz_{L} outputs from the Transformer block, we have to remove the x0x_{{0}}^{{}^{\prime}} vector from the zLz_{L}, because only then we can reshape it to the same size of FNF_{N}.

Refer to caption
Figure 4: (a) The visible light OCT device for capturing images. (b) Visible-light OCT B-scan image. (c) Annotation image (ground truth). (d) Average percentage of pixels on OCT images for each tissue layer besides the background(the percentage of background is 75.06%).

III-C MAA Module

To better preserve the semantic information at each encoder scale, the MAA module is designed in the skip connection part of the encoder and decoder by a large convolution kernel and channel attention mechanism. It’s also set with multi-scale feature fusion, and it’s based on the ideas of asymmetric convolution[15], multi-scale feature extraction[27, 28] and channel attention mechanism[29].

F0=Conv5×5(F0)F_{0}^{{}^{\prime}}=Conv_{5\times 5}\left(F_{0}\right) (4)
F0i=Convki×1[Conv1×ki(F0)],k=3,7,11F_{0i}^{{}^{\prime}}=Conv_{ki\times 1}\left[Conv_{1\times ki}\left(F_{0}^{{}^{\prime}}\right)\right],~{}k=3,7,11 (5)
F0i′′=CA(F0i),F_{0i}^{{}^{\prime\prime}}=CA\left(F_{0i}^{{}^{\prime}}\right), (6)

The workflow of the MAA module is illustrated in Fig. 3, it begins with the input of a feature map F0F_{0} extracted from the feature encoder, assuming a shape size of (CC, H0H_{0}, W0W_{0}). F0F_{0} enters MAA and goes through a 5×55\times 5 convolution kernel to expand the perceptual field, and then has to enter a multi-scale convolution to extract information at different scales (Eq. 4, Eq. 5). Each scale goes through a channel attention (CA) module to increase the weight of the feature map on the channel dimension (Eq. 6). As shown in the channel attention module in Fig. 3, the feature F0iC×H×WF_{0i}^{{}^{\prime}}\in\mathbb{R}^{C\times H\times W} reshape to F0i_cnC×NF_{0i\_cn}^{{}^{\prime}}\in\mathbb{R}^{C\times N}, where N=H×WN=H\times W, \in represents the shape of the feature map. After that we perform a matrix multiplication between F0i_cnF_{0i\_cn}^{{}^{\prime}} and the transpose of it, get the F0i_ccC×CF_{0i\_cc}^{{}^{\prime}}\in\mathbb{R}^{C\times C}. Eventually, we apply the softmax layer to get the channel attention map AC×CA\in\mathbb{R}^{C\times C}

F0i_cc=F0i_cn×(F0i_cn)TF_{0i\_cc}^{{}^{\prime}}=F_{0i\_cn}^{{}^{\prime}}\times\left(F_{0i\_cn}^{{}^{\prime}}\right)^{T} (7)
A=Softmax(F0i_cc)A=Softmax\left(F_{0i\_cc}^{{}^{\prime}}\right) (8)

(Eq. 7, Eq. 8). Here the matrix AC×CA\in\mathbb{R}^{C\times C} contains information about the weights between channels, e.g. (AC×C)ij\left(A\in\mathbb{R}^{C\times C}\right)_{ij} represents the impact of ithi^{th} channel on jthj^{th} channel. In addition, we multiply the matrix between the transpose of AC×CA\in\mathbb{R}^{C\times C} and

A1C×N=AT×F0i_cnA_{1}\in\mathbb{R}^{C\times N}=A^{T}\times F_{0i\_cn}^{{}^{\prime}}\\ (9)
F0i′′=αA1C×NreshapeC×H×W+F0iF_{0i}^{{}^{\prime\prime}}=\alpha A_{1}\in\mathbb{R}^{C\times N\overset{reshape}{\rightarrow}C\times H\times W}+F_{0i}^{{}^{\prime}} (10)

F0i_cnF_{0i\_cn}^{{}^{\prime}}, and reshape the result to C×H×W\mathbb{R}^{C\times H\times W}. Then we multiply the result by a scale parameter α\alpha and perform an element-wise sum operation with F0iF_{0i}^{{}^{\prime}} to obtain the final output F0i′′F_{0i}^{{}^{\prime\prime}} (Eq. 9 and Eq. 10). After the CA module, all feature maps are summed and convolved by 1×11\times 1 kernel to generate a weighted feature map, and we use the weighted feature map to re-weight F0F_{0} by element-wise multiplying (\bigotimes) it with the input F0F_{0} (Eq. 11 and Eq. 12).

att=Conv1×1[i=13F0i′′+CA(F0)]att=Conv_{1\times 1}\left[\sum_{i=1}^{3}F_{0i}^{{}^{\prime\prime}}+CA\left(F_{0}^{{}^{\prime}}\right)\right] (11)
FM0=attF0F_{M0}=att\bigotimes F_{0} (12)

At last, we obtain the FM0F_{M0}. According to the above procedure, we can input the extracted four scales feature maps F1F_{1}, …, FN1F_{N-1}, FNF_{N} into MAA module to get FM1F_{M1}, …, FMN1F_{MN-1}, FMNF_{MN}, respectively, as shown in Fig. 2.

Instead, the ASPP employs a series of dilated convolution layers with different rates to obtain larger receptive fields, however, dilated convolutions may lose local detail information. In contrast, the MAA module utilizes asymmetric convolutions with large kernels, which allows capturing multi-scale features without sacrificing too much detail. Furthermore, before the 1x1 convolution compresses the channels in the MAA module, a channel attention mechanism is adopted, enhancing the module’s focus on the channel dimension and better preserving important features.

TABLE I: Multiple approaches to multiple metrics (SmPAS_{mPA} (%), SmIoUS_{mIoU} (%), SDSCS_{DSC} (%), SPAS_{PA} (%)) evaluation on the Vis-105H dataset, the No. 1 and No. 2 places are bolded in black and bolded in gray for each metric, respectively.
Method Indicators Tissue Layers SmPAS_{mPA} SmIoUS_{mIoU}
NFL GCL INL/OPL ONL OS RPE
ReLayNet[11] SDSCS_{DSC} 93.6 85.6 85.5 93.5 87.9 94.1 97.3 82.1
SPAS_{PA} 93.3 82.1 86.0 97.2 87.5 91.1
OS_MGU[30] SDSCS_{DSC} 94.8 86.8 86.7 94.1 88.6 94.5 97.7 83.6
SPAS_{PA} 93.5 82.8 87.4 97.8 88.1 91.2
DFANet[31] SDSCS_{DSC} 91.1 80.0 80.8 88.3 73.5 88.8 95.1 72.6
SPAS_{PA} 91.3 77.0 82.9 93.3 74.4 86.3
BiSeNet[32] SDSCS_{DSC} 94.9 87.3 88.3 93.9 84.8 94.5 97.8 83.1
SPAS_{PA} 95.4 83.9 89.7 96.6 84.9 91.2
Attention_Unet[12] SDSCS_{DSC} 94.6 87.4 87.6 93.6 87.1 93.7 97.6 83.1
SPAS_{PA} 94.6 83.4 87.6 97.6 87.5 91.2
EMV-Net[22] SDSCS_{DSC} 94.6 85.3 85.5 93.7 87.9 94.5 97.6 82.5
SPAS_{PA} 93.5 81.9 86.2 97.1 90.1 91.7
U-net [10] SDSCS_{DSC} 92.6 82.5 83.4 91.5 82.6 92.3 96.8 78.1
SPAS_{PA} 90.3 80.8 84.8 96.6 80.8 90.0
SegFormer-B0 [20] SDSCS_{DSC} 94.5 83.4 84.1 93.4 88.2 94.2 97.4 81.6
SPAS_{PA} 92.8 79.0 86.9 97.6 86.3 91.3
TransUnet[18] SDSCS_{DSC} 94.4 86.9 87.5 94.1 87.5 94.7 97.7 83.5
SPAS_{PA} 93.4 83.2 88.0 97.0 87.2 91.6
LightReSeg (Our) SDSCS_{DSC} 94.3 87.5 89.1 94.7 89.1 94.5 97.8 84.6
SPAS_{PA} 94.3 84.5 89.4 96.9 87.8 91.6

III-D Light-weight Designs

The basic principle of our design of LightReSeg is to improve the overall performance of the model with as few parameters as possible, so we can meet the computational efficiency of deployments for real-time clinical applications. Thus, we make some light-weight designs for part of the structure of the model. For encoder extraction of multi-scale feature maps, we use depthwise separable convolutions (DS-Conv) for our down-sampling[14]. DS-Conv considers the channels and spatial regions of the feature map separately, takes different convolution kernels for different channels, and then uses point convolution to aggregate the channel information, achieving a richer feature representation with fewer parameters to learn. We find that using standard convolution for down-sampling is better than pooling, but convolution increases the number of parameters, so we use DS-Conv to ensure the segmentation performance while decreasing the burden of the number of parameters. In addition, as mentioned in Sec. III-C, multi-scale asymmetric convolution is used in our MAA module. Initially, our idea is to perform multi-scale extraction of feature information like ASPP [28], but we find that using standard convolution would cause a greater burden on the number of parameters, since the MAA module is supposed to perform extraction of the NN scale feature maps, which would certainly further increase the number of parameters. In order to reduce the number of parameters in the MAA module, we use asymmetric convolution instead of standard convolution, shown in Fig. 3, and the overall performance of LightReSeg is nearly the same. In particular, the Transformer block has a significant proportion of the number of parameters in LightReSeg. We have optimized the internal parameters of the Transformer block structure, such as the setting of heads=8 for multiple heads in MSA and dim_head=64 for hidden linear layers in MSA[13], etc. All these designs to minimize the number of parameters are made under the precondition of ensuring the performance of LightReSeg.

IV EXPERIMENT

In this section, we compare our LightReSeg against the state-of-the-art methods for retinal layer segmentation on i) two publicly available OCT datasets with unhealthy eyes, where the Glaucoma dataset is for the Glaucoma disease and the DME dataset is for Diabetic Macular Edema (DME), and ii) a new dataset, named as Vis-105H, that we create on top of raw images collected from [33] which correspond to only healthy human eyes with visible light OCT. Moreover, we present a thorough ablation study with our Vis-105H dataset to validate the effectiveness of the proposed MAA and the introduction of the Transformer block.

IV-A Datasets

Vis-105H dataset

We create the Vis-105H dataset for the method evaluation which contains images captured using a custom visible light OCT device that takes optic disc cubes from each subject [34, 35, 36], as shown in Fig. 4(a). The raw images are originally collected and presented in [33], and they correspond to 21 eyes from 14 healthy subjects, all of whom with written informed consent prior to the participation. The inclusion criteria include i) age 4040 years, ii) best corrected visual acuity better than 20/40, iii) no previous intraocular surgery of any kind, and iv) no known retinal disease, resulting in 21 scanning cubes with 5376 B-scan images. Two professional ophthalmologists further select 5 images with high imaging quality from every 256 B-scans, as shown in Fig. 4(b), with the selection criteria including i) low noise, ii) clear retinal layer boundaries, and iii) suitability for ophthalmologists to use for diagnosis. With the supervision of medical ophthalmology professionals, we annotate these 105 pre-processed retinal Visible-light OCT images of healthy human eyes for semantic segmentation, and complete the Vis-105H dataset for 7-class semantic segmentation including the background, as shown in Fig. 4(c). Specifically, we exploit the graphics software Inkscape to annotate the six layers, nerve fiber layer (NFL), ganglion cell layer (GCL) and inner plexiform layer (IPL), inner nuclear layer (INL) and out plexiform layer (OPL), outer nuclear layer (ONL), outer segment (OS), and retinal pigment epithelium (RPE), as red, yellow, green, light blue, dark blue, and pink, respectively, and to annotate the remaining regions as background black. Each annotated image is further examined by a professional ophthalmologist and exported in PNG format with a size of 660×\times300 pixels. Fig. 4(d) shows the average pixel percentages of all categories on the OCT image except for the background, from which we can observe that the most dominant two classes are the NFL and ONL layers while the OS and GCL layers are the least dominant classes. The Vis-105H dataset is divided into a training set, a validation set, and a test set consisting of 75, 15, and 15 images, respectively. The data for the Vis-105 dataset is conducted at BMC, and both BU and BMC hold ownership rights to the data.

Glaucoma dataset

The dataset is collected from 61 different subjects [30], where 12 radial OCT B scans of each subject are collected using DRI OCT-1 Atlantis at the Ophthalmology Department of Shanghai General Hospital. All images are scanned at 20.48 mm × 7.94 mm field of view in the optic nerve head region. Under the supervision of a glaucoma specialist, two ophthalmologists manually annotate these images into the optic disc and nine retinal layers. The image size is 1024 × 992 and the dataset is divided into training, validation, and test sets by 148, 48, and 48 images, respectively.

DME dataset

The dataset is collected by Chiu et al. using the Duke Enterprise Data Unified Content Explorer search engine to retrospectively identify DME subjects within the Duke Eye Center[37]. It consists of 110 OCT B-scans obtained from 10 patients with DME with a size of 496×\times768 pixels, each B-scan is annotated with 9 layers. The dataset is divided by 88, 11, 11 into training, validation, and test set, respectively. All subjects are scanned with the macula at the center.

TABLE II: Multiple approaches to multiple metrics (SmPAS_{mPA} (%), SmIoUS_{mIoU} (%), SDSCS_{DSC} (%), SPAS_{PA} (%)) evaluation on the DME dataset, the No. 1 and No. 2 places are bolded in black and bolded in gray for each metric, respectively.
Method Indicators Tissue Layers SmPAS_{mPA} SmIoUS_{mIoU}
NFL GCL/IPL INL OPL ONL/ISM ISE OS/RPE Fluid
ReLayNet[11] SDSCS_{DSC} 82.0 93.4 79.4 77.2 86.8 86.6 86.5 53.8 95.5 69.6
SPAS_{PA} 79.5 93.3 76.2 76.7 87.7 89.7 87.2 52.0
OS_MGU[30] SDSCS_{DSC} 82.6 93.6 79.5 78.4 87.2 87.1 85.6 57.7 95.6 69.8
SPAS_{PA} 80.7 93.7 77.0 77.2 88.3 89.5 84.6 52.1
DFANet[31] SDSCS_{DSC} 77.4 90.3 73.6 72.6 85.8 85.7 85.5 49.8 94.7 64.8
SPAS_{PA} 73.7 91.4 70.4 70.8 84.1 89.4 85.9 47.3
BiSeNet[32] SDSCS_{DSC} 80.9 92.7 77.9 75.3 86.3 86.1 86.6 56.7 95.4 68.3
SPAS_{PA} 76.7 93.0 76.8 72.2 86.9 87.5 89.8 51.0
Attention_Unet[12] SDSCS_{DSC} 80.5 91.5 77.6 75.9 86.8 86.6 86.4 55.6 95.3 68.0
SPAS_{PA} 79.4 92.4 74.4 74.0 86.8 88.9 86.6 49.9
EMV-Net[22] SDSCS_{DSC} 80.4 92.5 78.5 74.8 85.6 86.3 86.3 59.9 95.3 68.4
SPAS_{PA} 74.9 91.6 81.8 71.3 81.2 89.4 87.6 60.2
U-net [10] SDSCS_{DSC} 82.2 92.8 79.1 78.2 86.7 87.4 86.6 56.2 95.5 69.4
SPAS_{PA} 81.4 92.9 76.2 76.5 87.6 90.7 86.7 52.2
SegFormer-B0 [20] SDSCS_{DSC} 81.7 93.7 79.1 77.0 86.6 86.6 85.7 57.6 95.6 69.2
SPAS_{PA} 77.5 95.0 74.9 77.0 89.8 89.9 87.8 49.1
TransUnet[18] SDSCS_{DSC} 81.4 93.1 79.8 77.4 87.4 87.2 86.6 60.1 95.7 70.0
SPAS_{PA} 80.0 93.0 77.7 77.6 88.2 89.5 86.7 53.9
LightReSeg (Our) SDSCS_{DSC} 82.4 93.7 79.8 78.8 87.9 87.3 86.0 59.7 95.7 70.5
SPAS_{PA} 79.7 93.5 77.1 78.1 87.8 89.6 85.6 57.8
TABLE III: Multiple approaches to multiple metrics (SmPAS_{mPA} (%), SmIoUS_{mIoU} (%), SDSCS_{DSC} (%), SPAS_{PA} (%)) evaluation on the Glaucoma dataset, the No. 1 and No. 2 places are bolded in black and bolded in gray for each metric, respectively.
Method Indicators Tissue Layers SmPAS_{mPA} SmIoUS_{mIoU}
NFL GCL IPL INL OPL ONL IS/OS RPE Choroid OD
ReLayNet[11] SDSCS_{DSC} 79.4 64.9 70.8 76.4 81.1 90.5 86.0 93.1 87.9 77.8 94.2 67.0
SPAS_{PA} 77.2 60.7 70.6 77.0 80.5 90.0 84.9 85.2 88.6 77.4
OS-MGU[30] SDSCS_{DSC} 80.7 62.0 69.6 74.2 78.4 89.6 84.7 83.0 88.5 83.7 94.8 66.7
SPAS_{PA} 82.8 56.1 68.7 74.4 75.7 88.3 85.0 85.7 88.7 86.8
DFANet[31] SDSCS_{DSC} 79.8 60.5 68.2 74.3 77.4 89.4 85.0 81.3 88.2 83.9 94.7 65.8
SPAS_{PA} 80.0 55.9 66.1 71.9 73.1 91.7 85.7 80.8 89.6 88.6
BiSeNet[32] SDSCS_{DSC} 78.0 61.5 67.4 70.1 77.1 88.5 82.4 78.3 86.9 84.3 94.5 64.0
SPAS_{PA} 76.4 57.0 65.6 66.2 77.4 89.5 82.3 81.6 90.1 89.3
Attention_Unet[12] SDSCS_{DSC} 79.5 60.1 64.7 72.8 78.8 90.6 86.3 82.8 88.8 83.2 94.7 66.0
SPAS_{PA} 84.2 57.9 59.4 71.0 75.0 91.9 85.6 81.8 89.1 84.8
EMV-Net[22] SDSCS_{DSC} 81.2 65.3 69.5 72.1 79.1 90.0 83.9 81.5 88.5 83.4 94.8 66.6
SPAS_{PA} 83.7 59.3 67.8 65.8 74.8 88.6 81.5 83.4 89.9 86.8
U-net [10] SDSCS_{DSC} 81.9 63.7 69.4 73.5 77.5 89.7 85.4 82.3 87.8 83.3 94.8 66.6
SPAS_{PA} 83.5 58.1 68.3 73.7 74.6 91.7 84.2 83.4 91.0 86.5
SegFormer-B0 [20] SDSCS_{DSC} 82.1 64.3 69.4 76.4 76.6 89.9 84.4 81.2 88.6 84.2 94.8 66.9
SPAS_{PA} 86.7 58.5 63.9 81.7 69.7 92.6 82.8 79.8 87.7 90.7
TransUnet[18] SDSCS_{DSC} 80.6 60.0 69.4 76.1 80.2 90.8 85.8 82.6 88.5 82.4 94.7 67.0
SPAS_{PA} 84.7 53.3 69.0 77.6 78.5 93.6 83.0 80.3 86.9 84.3
LightReSeg (Our) SDSCS_{DSC} 80.9 64.3 70.5 77.2 80.4 90.1 85.4 82.7 88.4 83.3 94.9 67.8
SPAS_{PA} 83.5 61.1 70.8 78.8 78.7 90.2 85.0 85.3 90.5 83.5

IV-B Performance Metrics

We evaluate the segmentation performance quantitatively using the Dice similarity coefficient (DSC), mean Intersection over Union (mIoU), pixel accuracy (PA), and mean pixel accuracy (mPA). They can be computed as:

SDSC=2×TP2×TP+FP+FN,S_{DSC}=\frac{2\times TP}{2\times TP+FP+FN}, (13)
SmIoU=1k+1i=0kTPFN+FP+TP,S_{mIoU}=\frac{1}{k+1}\sum_{i=0}^{k}\frac{TP}{FN+FP+TP}, (14)
SPA=TP+TNTP+FP+TN+FNS_{PA}=\frac{TP+TN}{TP+FP+TN+FN} (15)

and

SmPA=1k+1i=0kTP+TNTP+FP+TN+FN,S_{mPA}=\frac{1}{k+1}\sum_{i=0}^{k}\frac{TP+TN}{TP+FP+TN+FN}, (16)

where TPTP, FPFP, TNTN, and FNFN represent True Positives, False Positives, True Negatives, and False Negatives respectively. kk represents the number of categories. In addition, we count the number of parameters for compared approaches to represent their computation complexity.

IV-C Implementation Details

LightReSeg is implemented based on the PyTorch and trained with the Adam optimizer with the cross-entropy loss. The initial learning rate is set to 0.001 which is then gradually halved every 40 epochs. Data augmentation is applied on all three datasets, including horizontal flipping with probability P=0.5, random center rotation within plus or minus 20 degrees with probability P=0.5, median and motion blur processing with P=0.5 probability, random Gaussian noise addition, as well as random brightness and contrast with P=0.5 probability. All experiments are conducted on an NVIDIA GeForce RTX 3090 Graphics card. The code of the proposed LightReSeg could be found at: https://github.com/Medical-Image-Analysis/LightReSeg.

IV-D Comparison

We compare LightReSeg with the state-of-the-art approaches including ReLaynet[11], EMV-Net[22], Attention_Unet[12], BiSeNet[32], DFANet[31], OS_MGU[30], U-net [10] and TransUnet[18]. We report the results on the above-mentioned three datasets.

IV-D1 Quantitative Analysis

Tab. I shows the comparison of our approach with state-of-the-art methods on the Vis-105H dataset. We can see that our approach LightReSeg scores the best performance in both SmPAS_{mPA} and SmIoUS_{mIoU} metrics, with +1%+1\% improvement in terms of the SmIoUS_{mIoU} compared to the second-best performing method OS_MGU [30]. In terms of the SDSCS_{DSC} metric, our approach achieves the highest segmentation performance in the GCL layer, INL/OPL layer, ONL layer, and OS layer. Regarding the SPAS_{PA} metric, our approach achieves the first place in the GCL layer, and the second place in the INL/OPL layer and RPE layer, which also shows the outstanding performance of our proposed approach. The segmentation performance of ReLayNet, the lightest algorithm in the field of retinal layer segmentation, is 2.5%-2.5\% lower than our proposed LightReSeg in terms of the SmIoUS_{mIoU} metric. We further perform a statistical significance test, using the Wilcoxon rank sum test, to compare the Dice Score performance of different methods on each layer. When comparing our approach with ReLayNet, we observe a PvalueP_{value} of 0.031250 (p<<0.05), indicating a statistically significant difference. The poor performance of ReLayNet might be due to its rather simple feature fusion strategy for combining features from encoder and decoder, i.e. skip connections, leading to limited semantic information at the fusion. Instead, LightReSeg further introduces the MAA module for the feature fusion at multiple scales. Moreover, we add Transformer layers at the bottom of the encoder for efficient global reasoning which further improves the segmentation accuracy. SegFormer is a new light-weight model in the field of semantic segmentation and has a good performance in many tasks [20], but as can be seen from Tab. I, SegFormer-B0 only ranks 7th in terms of the SmIoUS_{mIoU} and SmPAS_{mPA} metrics, while in terms of the SDSCS_{DSC} metric, its segmentation performance of all layers is much lower than ours. In terms of the SmPAS_{mPA} metric, only the ONL layer is higher than our by 0.7 percentage points, and the segmentation performance in the rest of the layers is much lower than our approach. We analyze that the poor segmentation performance of SegFormer-B0 on this task may be mainly due to two factors: first, its excessive pursuit of lightweight results in limited ability to extract detailed features; second, the up-sampling part of SegFormer-B0’s decoder is too rough, resulting in the lack of fine detail recovery.

Refer to caption
Figure 5: (a) and (b) are the performance of the nine mainstream approaches on the three retinal layer segmentation datasets measured by the mIoU and mPA metrics, respectively.

We evaluate our approach on the DME dataset, as shown in Tab. II. We achieve the best performance in terms of both SmPAS_{mPA} and SmIoUS_{mIoU}. In terms of SDSCS_{DSC} we achieve first place in 4 layers (i.e. GCL/IPL, INL, OPL, and ONL/ISM) and second place in another 2 layers (i.e. NFL and ISE), while in terms of SmPAS_{mPA}, we are on par with the state-of-the-art performance. Among other approaches, such as TransUnet, OS_MGU, and ReLayNet, although they are more advanced in the SmPAS_{mPA} metric, they are 0.5, 0.7, and 0.9 percentage points lower than the first place in the SmIoUS_{mIoU} metric, respectively, which also indicates that our model also has a strong comprehensive performance on the DME dataset. We further perform the statistical significance test by using the Wilcoxon rank sum test. For example, when comparing our method with TransUnet on the same domain, we observe a PvalueP_{value} of 0.017629 (p<<0.05), indicating a statistically significant difference. Similar statistically significant differences are observed, with PvaluesP_{values} of 0.023437 (p<<0.05), 0.039062 (p<<0.05), and 0.017755 (p<<0.05) respectively when comparing our method with OS_MGU, EMV-Net, and SegFormer-B0. We further evaluate our approach on the Glaucoma dataset, as shown in Tab. III. Our LightReSeg achieves the best results on both the overall metrics SmPAS_{mPA} and SmIoUS_{mIoU}. Additionally, our approach also scores better on most of the layers in terms of both SDSCS_{DSC} and SPAS_{PA}.

Fig. 5 (a) and (b) show the performance of the ten mainstream approaches on the three datasets measured by the mIoU and mPA metrics, respectively. Our approach LightReSeg achieve the best performance on all three datasets in terms of SmIoUS_{mIoU} and SmPAS_{mPA}. TransUnet performs is the second best-performing method with a small margin lower than ours on all three datasets, however, its number of parameters is 32 times of the one of our approach. The lightest model ReLayNet is inferior to our method in terms of performance by 2.5%-2.5\%, 0.9%-0.9\%, and 0.8%-0.8\% on the Vis-105H, DME, and Glaucoma dataset, respectively. Therefore, our approach achieves state-of-the-art performance with a relatively light-weight model.

In addition, we find that regardless of the method, certain layers (NFL, ONL, RPE) have significantly higher accuracy compared to other layers, as shown in Tab. I. We believe that class imbalance is the most likely cause, as the number of samples in a certain class is much larger than in other classes, resulting in the model learning better for that class during training and exhibiting higher accuracy in evaluation. From Fig. 4(d), we can see that the average pixel proportion of NFL, ONL, and RPE layers is relatively high, and Tab. I also shows that these three layers have higher accuracy compared to other layers. Also, there are differences in shape, color, texture, and other aspects among different layers, which may make certain categories of targets easier to segment while others are more difficult. In summary, we believe that the best way to solve this problem is to first address class imbalance.

Refer to caption
Figure 6: Comparison of segmentation prediction maps of mainstream approaches on the Vis-105H dataset. (a) Original image. (b) Ground truth (c) Prediction map of U-net. (d) Prediction map of DFANet. (e) Prediction map of TransUnet. (f) Prediction map of Attention_Unet. (g) Prediction map of EMV-Net. (h) Prediction map of ReLayNet. (i) Prediction map of SegFormer-B0. (j) Prediction map of LightReSeg. The black dashed line framed area in the image is the region where the boundary of the prediction layer is not smooth or inaccurate.
Refer to caption
Figure 7: Comparison of segmentation prediction maps of mainstream approaches on the Glaucoma dataset. (a) Original image. (b) Ground truth (c) Prediction map of OS_MGU. (d) Prediction map of DFANet. (e) Prediction map of TransUnet. (f) Prediction map of Attention_Unet. (g) Prediction map of EMV-Net. (h) Prediction map of ReLayNet. (i) Prediction map of SegFormer-B0. (j) Prediction map of LightReSeg. The black dashed line framed area in the image is the region where the boundary of the prediction layer is not smooth or inaccurate, the red dashed area is where the prediction map has a segmentation category error.

IV-D2 Qualitative Analysis

Fig. 6 shows the segmentation prediction maps of several mainstream segmentation methods in a retinal B-scan image that contains blood flow information disturbance. As shown in Fig. 6(j), our approach achieves much better segmentation performance. The prediction graphs of all the approaches in Fig. 6(c), (d), (f), (h) all showed false positive errors, specifically, the region above the NFL layer or below the RPE layer of the retina is incorrectly identified as the retinal layer and segmented, resulting in the wrong segmentation target. The TransUnet in Fig. 6(e) performs well, but two locations on the boundary between the NFL and GCL layers are inaccurately segmented, as marked by the two locations in Fig. 6(e). The same occurs in EMV-Net in one place, as shown in Fig. 6(g). In comparison, our approach is not only without false positive errors, but also segmentation boundaries are more accurate.

Fig. 7 shows the comparative prediction plots of several mainstream segmentation approaches in a single B-scan image over the retinal optic nerve head region. As shown in Fig. 7(j), our approach shows a better segmentation performance. The black dashed region in Fig. 7(d), where the NFL and GCL layer boundaries appear distinctly jagged, and the black dashed region in Fig. 7(f), where a distinct honeycomb shape appears, are signs of unsmooth and inaccurate layer boundary prediction, and the same occurs in Fig. 7(c), (e), and (h). In addition, the problem of segmentation category error occurs, such as the red dashed area in Fig. 7(d), where the segmentation results in incorrectly segmenting part of the OPL layer into ONL layers, and this segmentation category error occurs at least twice in Fig. 7, (e), (f), and (g). In contrast, our approach not only has no problem with segmentation category error but also the segmented boundary is smoother than other approaches.

IV-E Ablation study

IV-E1 Performance of Different Modules

We conduct ablation experiments on the Vis-105H dataset with the main purpose of verifying the effect of the MAA module and the contribution of the Transformer block on the overall method segmentation performance. The ablated methods are: Base is the U-shaped backbone; Base_MAA and Base_trans3 are MAA and 3 transformer layers added to Base, respectively; Base_MAA_trans3 and Base_MAA_trans6 are based on Base_MAA with 3 and 6 transformer layers respectively. Note that Base_MAA_trans3 represents the final version of our proposed LightReSeg. As shown in Tab. IV, the Base_MAA approach improves over the Base approach, the SmPAS_{mPA} and SmIoUS_{mIoU} by +0.4%+0.4\% and +1.9%+1.9\%, respectively. All six layers except for the OS layer are also significantly improved in the SDSCS_{DSC}. We also observe that Base_MAA_trans3 improves over Base_trans3 in terms of both SmPAS_{mPA} and SmIoUS_{mIoU}. Fig. 8(c) and (d) show that the segmentation performance of the model is greatly improved by adding the MAA module to the model, and the results of mis-segmentation are significantly reduced in the background region below the RPE layer.

Regarding the Transformer block, as shown in Tab. IV, Base_trans3 improves +0.5%+0.5\% and +2.4%+2.4\% over Base in terms of SmPAS_{mPA} and SmIoUS_{mIoU}, respectively, and the segmentation ability of each layer is improved in terms of SDSCS_{DSC}. In addition, we also find that increasing the number of Transformer layers can further improve the segmentation performance within a certain range. For example, Base_MAA_trans6 adds 3 more Transformer layers than Base_MAA_trans3, which improves SmPAS_{mPA} and SmIoUS_{mIoU} by +0.1%+0.1\% and 0.6%0.6\%, respectively, this comes at a cost of +40%+40\% more of parameters. Fig. 8(c) and (e) show qualitatively that adding three sequential Transformer blocks to the model contributes to a big improvement in the segmentation performance of the model, and there is no segmentation error in the background region below the RPE layer.

TABLE IV: The proposed approach performs multiple metrics evaluation of ablation experiments on the Vis-105H dataset.
Method SDSCS_{DSC} (%) Results of Tissue Layers SmPAS_{mPA} (%) SmIoUS_{mIoU} (%) Params (M)
NFL GCL INL/OPL ONL OS RPE
Base 94.0 84.6 84.8 93.2 87.9 92.6 97.2 81.3 1.61
Base_MAA 94.4 86.9 87.7 94.0 87.4 94.0 97.6 83.2 1.72
Base_trans3 94.7 86.9 87.3 94.1 89.0 94.0 97.7 83.7 3.20
Base_MAA_trans3(Our) 94.3 87.5 89.1 94.7 89.1 94.5 97.8 84.6 3.30
Base_MAA_trans6 95.0 87.6 89.0 94.8 89.7 95.1 97.9 85.2 4.68
Refer to caption
Figure 8: Prediction images of our method under different ablation settings. (a) Original image. (b) Ground truth. (c) Base framework (d) Base framework plus MAA module (e) Base framework plus three Transformer blocks (d) Base framework plus MAA module and six Transformer blocks.

To further interpret the role of different modules, we use a modified version of Grad-CAM[38] to visualize the feature activations in different layers of the model (see Fig. 9). Specifically, in accordance with the model settings in Tab. IV, we extract the feature activation heat maps for different retinal layers at different model structures and at different positions of the model, respectively. By comparing Fig. 9(e) and (i), with the addition of the MAA module, we can see that the heat map of the corresponding NFL layer in Fig. 9(i) is further enhanced compared to Fig. 9(e). The same trend is also observed in Fig. 9(h) and (l). It is further found that with the addition of the MAA module to the Base_trans3 model structure, the feature region corresponding to the segmentation category is also further enhanced, as shown in Fig. 9(p) and (t). Adding the Transformer blocks to the model also causes changes in the heat map, as in Fig. 9(f) and (n). The heat map of the corresponding RPE layer region is activated more comprehensively with the addition of the Transformer blocks, and the OS and background layers of the adjacent regions also show different degrees of activation.

Refer to caption
Figure 9: Heatmap for different retinal layers at different model structures and at different positions of the model, (a–d) for the single category prediction, (e–h) for the Base structure, (i–l) for the Base_MAA structure, (m–p) for the Base_trans3 structure, and (q–t) for the Base_MAA_trans3 (Our) structure. The white line highlights the contrast area in the graph.

When analyzing the activation maps, we observe that not only the regions corresponding to the predicted layers prominently highlighted, but also the feature weights of the adjacent layers are also significantly amplified. To interpret this behavior, we conduct further analysis and find that the boundaries of the retinal layers exhibit a high degree of correlation with the adjacent layers. As in Fig. 9(e), the focus is mainly on the NFL layer, but adjacent layers (e.g., background and GCL) also show various degrees of activation. Similarly, in Fig. 9(r), the RPE layer of the retinal layer is the main focus, while information from the OS and background layers at the upper and lower boundaries of the retinal layer is also utilized. This phenomenon suggests that the model uses information from adjacent layers to enhance its predictions. In addition, we also export the feature activation maps for different deep network layers in the prediction of the RPE layer, and we find that the heat map of the deep network focuses on more detailed and accurate regions compared to the heat map of the superficial network. As in Fig. 9(o), the regions where the heat map is focused on activation are highly coincident with the RPE layer, while the regions activated by the heat map in Fig. 9(n) are much coarser, which is similarly shown in Fig. 9(r) and (s). This observation suggests that the deeper extract image features of the model network contain more accurate and insightful semantic information. Overall, the visualization of deep feature activation provides valuable insights into the model’s decision process, and the model can utilize multiple layers of information to obtain more robust segmentation results.

IV-E2 Light-weight Setup

To evaluate the potential impact of DS-Conv on feature extraction capabilities, we replace all DS-Conv in the encoder section with standard convolutions and conduct comparative analyses[39], as shown in Tab. V. The results indicate that using standard convolutions instead of DS-Conv does not enhance feature extraction in our optimized model and leads to an 11% increase in the parameter count. This demonstrates that a simple increase in the model’s parameter count does not always result in performance.

TABLE V: Impact of DS-Conv on Model Performance.
Method Settings SmPAS_{mPA}(%) SmIoUS_{mIoU}(%) Params(M)
Encoder with DS-Conv 97.8 84.6 3.30
Encoder with Conv 97.8 84.3 3.69

For assessing the impact of model light-weight on performance, we modify the number of channels in the encoder’s multi-scale features, both halving and doubling them, the specific results are shown in Tab. VI. We find that, based on our model, when the channel of the multi-scale features extracted by the encoder is reduced to 0.5 and 0.25 times, the model’s parameter count decreased, but its segmentation performance significantly deteriorated. When we increase the channel count to 2 and 4 times, there is no significant change in the model’s accuracy, yet the parameter counts are 3 and 11 times the original size, respectively. From this, we can also conclude that our lightweight design maintains accuracy while using the fewest possible parameters.

TABLE VI: Impact of Model Light-weight on Performance.
Method Settings mPA(%) mIoU(%) Params(M)
0.25x 96.9 79.2 0.48
0.5x 96.9 79.1 1.18
1x 97.8 84.6 3.30
2x 97.8 84.3 9.87
4x 97.7 84.1 33.48

IV-F Limitation

Although LightReSeg boasts only 3.3M trainable parameters and delivers the optimal retinal layer segmentation outcomes, in real-world applications, model parameter size is not the sole determinant of algorithm efficacy. Inference speed emerges as a crucial metric for assessing algorithmic effectiveness. As evidenced by Tab. VII, our proposed method exhibits inference speeds of 0.11ss, 0.27ss, and 0.07ss on the three datasets, respectively. While these figures already signify remarkable efficiency, there remains scope for further enhancement. For example, although the segmentation accuracy of ReLayNet is far inferior to our approach, it must be admitted that its inference efficiency is slightly higher than ours.

TABLE VII: Statistical inference speed of different methods on three datasets, and the best performance on each dataset is bolded.
Method
Params
Vis-105H
DME
Glaucoma
ReLayNet[11] 0.7MM 0.08ss 0.20ss 0.06ss
EMV-Net[22] 1.9MM 0.15ss 0.31ss 0.13ss
OS_MGU[30] 2.0MM 0.12ss 0.30ss 0.07ss
DFANet[31] 2.1MM 0.12ss 0.31ss 0.08ss
SegFormer-B0 [20] 3.7MM 0.12ss 0.28ss 0.07ss
BiSeNet[32] 13.1MM 0.08ss 0.20ss 0.05ss
U-net [10] 34.5MM 0.10ss 0.20ss 0.08ss
Attention_Unet[12] 34.8MM 0.10ss 0.22ss 0.09ss
TransUnet[18] 105.7MM 0.14ss 0.32ss 0.11ss
LightReSeg (Our) 3.3MM 0.11ss 0.27ss 0.07ss

Moreover, in the context of practical clinical applications, constraints related to dataset limitations preclude the inclusion of all device data types. Imagery obtained through disparate devices may exhibit domain discrepancies, inevitably resulting in false positive errors in the segmentation output for unfamiliar new devices. Consequently, apart from enlarging the dataset, enhancing the model’s domain generalization competency is a subject worthy of further investigation. The supplementary segmentation result evaluation system will also augment ophthalmologists’ confidence in the segmentation outcomes.

V CONCLUSION

In this paper, we propose a novel light-weight method LightReSeg for retinal layer segmentation. Our method introduces a Transformer-based block in the encoder part to enable global reasoning for reducing errors in the background region and an attention mechanism, named MAA module, to best exploit rich semantic information for the multi-scale feature fusion to improve the segmentation accuracy. The extensive evaluation shows that our approach achieves the best segmentation performance on both our collected Vis-105H dataset and two other public ones, i.e. DME and Glaucoma, indicating that our model has good reliability in the face of noise and uncertainty in the data. To improve the efficiency, our method also incorporates light-weight designs. The ablation experiments detailed in Section IV-E2 demonstrate that our lightweight design is precisely adequate, maintaining the highest level of accuracy while significantly reducing the number of parameters compared to other high-performing methods. This further highlights the practicality of our proposed approach, particularly evident in the real-time preview functionality of small-field-of-view OCTA imaging. In the future, we will collect more clinical data to further improve our approach and continuously contribute to the efficient performance of OCT devices.

References

  • [1] F. K. Horn, C. Y. Mardin, R. Laemmer, D. Baleanu, A. M. Juenemann, F. E. Kruse, and R. P. Tornow, “Correlation between local glaucomatous visual field defects and loss of nerve fiber layer thickness measured with polarimetry and spectral domain oct,” Investigative ophthalmology & visual science, vol. 50, no. 5, pp. 1971–1977, 2009.
  • [2] F. Pollet-Villard, C. Chiquet, J.-P. Romanet, C. Noel, and F. Aptel, “Structure-function relationships with spectral-domain optical coherence tomography retinal nerve fiber layer and optic nerve head measurements,” Investigative ophthalmology & visual science, vol. 55, no. 5, pp. 2953–2962, 2014.
  • [3] A. Ajaz, H. Kumar, and D. Kumar, “A review of methods for automatic detection of macular edema,” Biomedical Signal Processing and Control, vol. 69, p. 102858, 2021.
  • [4] C. Brandl, C. Brücklmayer, F. Günther, M. E. Zimmermann, H. Küchenhoff, H. Helbig, B. H. Weber, I. M. Heid, and K. J. Stark, “Retinal layer thicknesses in early age-related macular degeneration: results from the german augur study,” Investigative ophthalmology & visual science, vol. 60, no. 5, pp. 1581–1594, 2019.
  • [5] J. A. van de Kreeke, H.-T. Nguyen, J. den Haan, E. Konijnenberg, J. Tomassen, A. den Braber, M. Ten Kate, L. Collij, M. Yaqub, B. van Berckel et al., “Retinal layer thickness in preclinical alzheimer’s disease,” Acta ophthalmologica, vol. 97, no. 8, pp. 798–804, 2019.
  • [6] D. Huang, E. A. Swanson, C. P. Lin, J. S. Schuman, W. G. Stinson, W. Chang, M. R. Hee, T. Flotte, K. Gregory, and C. A. Puliafito, “Optical coherence tomography,” Science, vol. 254, no. 5035, pp. 1178–1181, 1991.
  • [7] X. Shu, L. J. Beckmann, and H. F. Zhang, “Visible-light optical coherence tomography: a review,” Journal of biomedical optics, vol. 22, no. 12, p. 121707, 2017.
  • [8] W. Song, W. Shao, and J. Yi, “Wide-field and micron-resolution visible light optical coherence tomography in human retina by a linear-k spectrometer,” in Bio-Optics: Design and Application.   Optica Publishing Group, 2021, pp. DM2A–4.
  • [9] W. Song, W. Shao, W. Yi, and J. Yi, “Linear k-domain wide-field and micron-resolution visible light human retinal optical coherence tomography,” in Ophthalmic Technologies XXXII.   SPIE, 2022, p. PC1194102.
  • [10] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention.   Springer, 2015, pp. 234–241.
  • [11] A. G. Roy, S. Conjeti, S. P. K. Karri, D. Sheet, A. Katouzian, C. Wachinger, and N. Navab, “Relaynet: retinal layer and fluid segmentation of macular optical coherence tomography using fully convolutional networks,” Biomedical optics express, vol. 8, no. 8, pp. 3627–3642, 2017.
  • [12] O. Oktay, J. Schlemper, L. L. Folgoc, M. Lee, M. Heinrich, K. Misawa, K. Mori, S. McDonagh, N. Y. Hammerla, B. Kainz et al., “Attention u-net: Learning where to look for the pancreas,” arXiv preprint arXiv:1804.03999, 2018.
  • [13] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  • [14] F. Chollet, “Xception: Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1251–1258.
  • [15] X. Ding, Y. Guo, G. Ding, and J. Han, “Acnet: Strengthening the kernel skeletons for powerful cnn via asymmetric convolution blocks,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 1911–1920.
  • [16] S. Wang, L. Yu, X. Yang, C.-W. Fu, and P.-A. Heng, “Patch-based output space adversarial learning for joint optic disc and cup segmentation,” IEEE transactions on medical imaging, vol. 38, no. 11, pp. 2485–2495, 2019.
  • [17] J. Li, G. Gao, L. Yang, G. Bian, and Y. Liu, “Dpf-net: A dual-path progressive fusion network for retinal vessel segmentation,” IEEE Transactions on Instrumentation and Measurement, 2023.
  • [18] J. Chen, Y. Lu, Q. Yu, X. Luo, E. Adeli, Y. Wang, L. Lu, A. L. Yuille, and Y. Zhou, “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021.
  • [19] M. Astaraki, Ö. Smedby, and C. Wang, “Prior-aware autoencoders for lung pathology segmentation,” Medical Image Analysis, p. 102491, 2022.
  • [20] E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, “Segformer: Simple and efficient design for semantic segmentation with transformers,” Advances in neural information processing systems, vol. 34, pp. 12 077–12 090, 2021.
  • [21] B. Chen, T. Niu, W. Yu, R. Zhang, Z. Wang, and B. Li, “A-net: An a-shape lightweight neural network for real-time surface defect segmentation,” IEEE Transactions on Instrumentation and Measurement, 2023.
  • [22] X. He, Y. Wang, F. Poiesi, W. Song, Q. Xu, Z. Feng, and Y. Wan, “Exploiting multi-granularity visual features for retinal layer segmentation in human eyes,” Frontiers in Bioengineering and Biotechnology, vol. 11, p. 1191803, 2023.
  • [23] B. Wang, W. Wei, S. Qiu, S. Wang, D. Li, and H. He, “Boundary aware u-net for retinal layers segmentation in optical coherence tomography images,” IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 8, pp. 3029–3040, 2021.
  • [24] X. Liu, J. Cao, S. Wang, Y. Zhang, and M. Wang, “Confidence-guided topology-preserving layer segmentation for optical coherence tomography images with focus-column module,” IEEE Transactions on Instrumentation and Measurement, vol. 70, pp. 1–12, 2020.
  • [25] K. Hu, D. Liu, Z. Chen, X. Li, Y. Zhang, and X. Gao, “Embedded residual recurrent network and graph search for the segmentation of retinal layer boundaries in optical coherence tomography,” IEEE Transactions on Instrumentation and Measurement, vol. 70, pp. 1–17, 2021.
  • [26] M. Gende, V. Mallen, J. de Moura, B. Cordón, E. Garcia-Martin, C. I. Sánchez, J. Novo, and M. Ortega, “Automatic segmentation of retinal layers in multiple neurodegenerative disorder scenarios,” IEEE Journal of Biomedical and Health Informatics, 2023.
  • [27] M.-H. Guo, C.-Z. Lu, Q. Hou, Z. Liu, M.-M. Cheng, and S.-M. Hu, “Segnext: Rethinking convolutional attention design for semantic segmentation,” arXiv preprint arXiv:2209.08575, 2022.
  • [28] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 834–848, 2017.
  • [29] J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, and H. Lu, “Dual attention network for scene segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 3146–3154.
  • [30] J. Li, P. Jin, J. Zhu, H. Zou, X. Xu, M. Tang, M. Zhou, Y. Gan, J. He, Y. Ling, and Y. Su, “Multi-scale gcn-assisted two-stage network for joint segmentation of retinal layers and discs in peripapillary oct images,” Biomed. Opt. Express, vol. 12, no. 4, pp. 2204–2220, Apr 2021. [Online]. Available: http://opg.optica.org/boe/abstract.cfm?URI=boe-12-4-2204
  • [31] H. Li, P. Xiong, H. Fan, and J. Sun, “Dfanet: Deep feature aggregation for real-time semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 9522–9531.
  • [32] C. Yu, J. Wang, C. Peng, C. Gao, G. Yu, and N. Sang, “Bisenet: Bilateral segmentation network for real-time semantic segmentation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 325–341.
  • [33] W. Song, S. Zhang, Y. M. Kim, N. Sadlak, M. G. Fiorello, M. Desai, and J. Yi, “Visible light optical coherence tomography of peripapillary retinal nerve fiber layer reflectivity in glaucoma,” Translational vision science & technology, vol. 11, no. 9, pp. 28–28, 2022.
  • [34] W. Song, S. Zhang, N. Sadlak, M. G. Fiorello, M. Desai, and J. Yi, “Visible and near-infrared optical coherence tomography (vnoct) in normal, suspect, and glaucomatous eyes,” in Ophthalmic Technologies XXXI, vol. 11623.   SPIE, 2021, p. 1162311.
  • [35] J. Yi and W. Song, “Systems and methods for fiber-based visible and near infrared optical coherence tomography,” Aug. 4 2020, uS Patent 10,732,354.
  • [36] W. Song, L. Zhou, S. Zhang, S. Ness, M. Desai, and J. Yi, “Fiber-based visible and near infrared optical coherence tomography (vnoct) enables quantitative elastic light scattering spectroscopy in human retina,” Biomedical Optics Express, vol. 9, no. 7, pp. 3464–3480, 2018.
  • [37] S. J. Chiu, M. J. Allingham, P. S. Mettu, S. W. Cousins, J. A. Izatt, and S. Farsiu, “Kernel regression based segmentation of optical coherence tomography images with diabetic macular edema,” Biomedical optics express, vol. 6, no. 4, pp. 1172–1194, 2015.
  • [38] K. Vinogradova, A. Dibrov, and G. Myers, “Towards interpretable semantic segmentation via gradient-weighted class activation mapping (student abstract),” in Proceedings of the AAAI conference on artificial intelligence, vol. 34, no. 10, 2020, pp. 13 943–13 944.
  • [39] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.