Light-weight Retinal Layer Segmentation with Global Reasoning
Abstract
Automatic retinal layer segmentation with medical images, such as optical coherence tomography (OCT) images, serves as an important tool for diagnosing ophthalmic diseases. However, it is challenging to achieve accurate segmentation due to low contrast and blood flow noises presented in the images. In addition, the algorithm should be light-weight to be deployed for practical clinical applications. Therefore, it is desired to design a light-weight network with high performance for retinal layer segmentation. In this paper, we propose LightReSeg for retinal layer segmentation which can be applied to OCT images. Specifically, our approach follows an encoder-decoder structure, where the encoder part employs multi-scale feature extraction and a Transformer block for fully exploiting the semantic information of feature maps at all scales and making the features have better global reasoning capabilities, while the decoder part, we design a multi-scale asymmetric attention (MAA) module for preserving the semantic information at each encoder scale. The experiments show that our approach achieves a better segmentation performance compared to the current state-of-the-art method TransUnet with 105.7M parameters on both our collected dataset and two other public datasets, with only 3.3M parameters.
Index Terms:
retinal layer segmentation, light-weight, multi-scale asymmetric attention, visible-light OCT.I INTRODUCTION
With the accelerated pace of lifestyle and the popularity of electronic products, many people lack awareness of eye protection, leading to an increasing incidence of retinal diseases. Many retinal diseases are accompanied by changes in the thickness of the retinal layers. For instance, in patients with glaucoma, the retinal nerve fiber layer is thinner and the optic disc is atrophied compared to the healthy eyes [1, 2]. In diabetic retinopathy, the macular region of the retina thickens and becomes edematous [3]. Current research indicates that the changes in the retinal layer begin to occur in the early stages of the disease[4, 5], therefore, automatic analysis of the retinal layers and monitoring their morphological changes can help to understand the disease progression and provide early treatment.
OCT can be used for non-invasive 3D structural imaging of the retina [6], with which experienced ophthalmologists can manually inspect the retinal layers to identify the disease progression. However, manual inspection is inefficient and tedious. It is also challenging for various levels of ophthalmologists to ensure consistency and objectivity of the inspection. Although imaging techniques with a higher resolution are available now, such as Visible-light OCT images [7, 8, 9], this does not address completely the above status quo. Recent convolutional neural networks (CNNs), as typified by U-net [10], have achieved promising results regarding semantic segmentation in the field of medical image analysis. The seminar work ReLayNet [11] has triggered widespread interest in applying CNNs to segment retinal layers. ReLayNet follows a multi-scale encoder-decoder structure, which has inspired many mainstream semantic segmentation frameworks to adopt, such as U-net [10] and Attention_Unet [12]. However, such structure can be limited in two main aspects. Firstly, most U-shape encoder-decoder structures perform the multi-scale feature extraction followed by feature fusion via only residual connections. Thus, it is limited to maintain and fully exploit the semantic information of feature maps at all scales, resulting in false positives in the background region when performing retinal layer segmentation. Secondly, many of these methods do not pay attention to the actual application needs of OCT devices, often pursuing accuracy unilaterally, for example, such multi-scale reasoning often comes at the cost of a large number of parameters, which can be computationally demanding for the deployment of real-time clinical applications.
To overcome the above-mentioned limitations, we propose LightReSeg, a novel multi-scale encoder-decoder network for end-to-end retinal layer segmentation, which is a light-weight segmentation approach tailored to practical OCT device requirements. In order to reduce segmentation errors in the background region, we propose to incorporate a Transformer block [13] to features with global reasoning and to introduce a multi-scale asymmetric attention (MAA) module to better preserve the semantic information at each encoder scale. To achieve computation efficiency, both the backbone and the MAA module adopt light-weight feature extractors, such as depthwise separable convolution and asymmetric convolution [14, 15]. LightReSeg achieves the state-of-the-art segmentation accuracy with only 3.3M parameters, a significantly light model, as shown in Fig. 1.

In summary, our contributions are:
-
•
We propose LightReSeg, a novel encoder-decoder structure for retinal layer segmentation, which exploits global information in a light-weight manner, to improve the segmentation accuracy with reduced computational complexity, it provides valuable experience for algorithms designed for practical applications in medical devices.
-
•
We propose a novel attention module, the MAA module, to jointly work with Transformer, in order to address erroneous segmentation in the background area by allowing the model to reason in a global manner.
-
•
We perform our method on the visible-light OCT images for the first time and score the best segmentation performance, it provides experience for the algorithm to perform robustly on datasets from different domains.
This paper is organized as follows: Sec. II describes the research progress in the field of biological tissue segmentation and retinal layer segmentation. Sec. III introduces the overall framework, the feature extractor, MAA module and the light-weight designs of the approach. Sec. IV describes the datasets, performance metrics, implementation details, analyzes quantitatively and qualitatively the experimental results, and performs ablation experiments. Finally, the conclusions of the paper are given in Sec. V.
II RELATED WORK
In this section, we first provide an overview of the research progress of semantic segmentation, especially light-weight semantic segmentation methods, followed by a more specific coverage of deep-learning-based retinal layer segmentation.
II-A Medical Image Segmentation
Medical image segmentation has made rapid progress under the promotion of deep learning, especially in the fields of biological tissue recognition and lesion detection, which has sparked a large amount of research on medical image segmentation. The morphology of the optic disc and optic cup, as well as the cup-to-disc ratio, can be used to assess glaucoma, Wang et al. proposed an automatic segmentation approach based on CNNs to accurately segment the optic disc and optic cup from fundus images for glaucoma detection[16]. Precise segmentation of retinal vessels from fundus images is essential for intervention in numerous diseases and high-precision segmentation of retinal vessels still remains a challenging task, Li et al. proposed a dual-path progressive fusion network, named DPF-Net, which achieved better segmentation ability by effectively fusing global and local features[17]. Abdominal multi-organ image segmentation plays a crucial role in the diagnosis and treatment of many diseases. traditional methods of manual depiction are inefficient and subjective, therefore, Chen et al. proposed TransUnet based on CNN and Transformer block for abdominal multi-organ segmentation[18]. TransUnet was the first model to introduce the Transformer block into the U-shaped encoder-decoder structure and achieved very high accuracy in segmentation. There are also segmentation studies dedicated to lesion regions, such as Astaraki et al. who segmented different types of lung pathological regions to enable screening for lung diseases[19]. In addition, some lightweight segmentation methods for practical applications have also been proposed, such as SegFormer[20], A-net[21], EMV-Net[22].

II-B Retinal Layer Segmentation
The morphology of the retinal layer is associated with the development of many ophthalmic diseases, therefore, there are many studies on retinal layer segmentation based on CNN and OCT images. The success of ReLayNet opened the door to the application of CNNs for retinal layer segmentation[11], a typical U-shaped network based on an encoder-decoder structure, which achieved the best performance for retinal layer segmentation at that time. Since layer segmentation of OCT retinal images is prone to speckle noise, intensity inhomogeneity, Wang et al. proposed a boundary-aware U-Net for retinal layer segmentation by detecting accurate boundaries[23]. To solve the topological errors caused by not considering the order of retinal layers, Liu et al proposed a novel deep learning-based framework that employed the distance maps of layer surfaces to convert the layer segmentation task into a multitasking problem for classification and regression[24]. To ensure the continuity of the retinal layer boundary and obtain more accurate segmentation results, Hu et al proposed a coarse-to-fine retinal layer boundary segmentation method based on the embedded residual recurrent network and the graph search, it has achieved good performance in both qualitative and quantitative indicators[25]. In addition, considering the impact of different neurological diseases on the retinal layer, Gende et al presented a fully automatic approach for the retinal layer segmentation in multiple neurodegenerative disorder scenarios[26], the results indicate that the model is more robust. There are also studies aimed at practical application deployment, such as He et al. proposed a light-weight retinal layer segmentation approach that achieved good performance with a lower number of parameters[22].
III METHODOLOGY
III-A Framework Overview
Our proposed LightReSeg is a U-shape network based on an encoder-decoder structure as shown in Fig. 2. The network takes a retinal layer OCT image as input. The first scale feature maps (, …, , ) are extracted by the encoder which consists of a multi-scale feature extractor, respectively. In order to further deepen the global reasoning capability in the depth direction, the feature map , which has to pass through a depth feature encoder: Transformer block[13]. The feature map further encoded by the Transformer block has no change in shape and size, but it facilitates long-range global reasoning. To better preserve details at different scales, multi-scale features (, …, , ) are then merged with its corresponding layers in the decoder part via our proposed MAA module, which outputs the feature maps (, …, , ) with the same shape size as its input. The output from the Transformer block is merged with the for feature fusion, and then after 2 fold size up-sampling, it continues to fuse with for feature fusion and then follows a similar operation until the fusion of the up-sampled restored feature map with is completed, and after channel adjustment, the final retinal layer segmentation image is output. LightReSeg is designed for real-time clinical operations, where light-weight designs are employed where possible.
III-B Multi-scale Encoder
The multi-scale encoder extracts scale features to fully exploit the semantic information of feature maps at all scales, further, in order to make the features have better global reasoning capabilities, we incorporate a Transformer block[13], which is a module that further optimizes the th scale feature map extracted from the deepest scale of the encoder. To ensure the dimensionality is consistent with the Transformer block port, we pre-process the feature map before being processed by the Transformer block. The feature map is first converted into a patch sequence as
(1) |
where is the number of channels in the feature map, is the size of the patch, and and are the height and width of the feature map.
We use a trainable linear projection to map the vector patch to a latent D-dimensional embedding space. To encode the spatial information of the patches, we learn specific position embeddings that are added to preserve the positional information in the patch embedding as
(2) |
where represents the patch embedding projection, the is an additional learnable vector concatenated with the remaining vectors to integrate the information of all remaining vectors, the is the concatenation operator, and represents the position embedding[13].

Then, inputting into the Transformer Layer, as shown in Fig. 2, first goes through the first Layer Norm (LN) layer and then enters the Multi-head Self-attention (MSA) layer, here forms the residual structure and get the . Then, after passing through the second LN layer, it enters the Multi-Layer Perceptron (MLP) layer, here once again forms the residual structure and get the .
At this point, the first Transformer layer is finished. We have set three Transformer layers in our approach, so we have to cycle three times.
(3) |
When the outputs from the Transformer block, we have to remove the vector from the , because only then we can reshape it to the same size of .

III-C MAA Module
To better preserve the semantic information at each encoder scale, the MAA module is designed in the skip connection part of the encoder and decoder by a large convolution kernel and channel attention mechanism. It’s also set with multi-scale feature fusion, and it’s based on the ideas of asymmetric convolution[15], multi-scale feature extraction[27, 28] and channel attention mechanism[29].
(4) |
(5) |
(6) |
The workflow of the MAA module is illustrated in Fig. 3, it begins with the input of a feature map extracted from the feature encoder, assuming a shape size of (, , ). enters MAA and goes through a convolution kernel to expand the perceptual field, and then has to enter a multi-scale convolution to extract information at different scales (Eq. 4, Eq. 5). Each scale goes through a channel attention (CA) module to increase the weight of the feature map on the channel dimension (Eq. 6). As shown in the channel attention module in Fig. 3, the feature reshape to , where , represents the shape of the feature map. After that we perform a matrix multiplication between and the transpose of it, get the . Eventually, we apply the softmax layer to get the channel attention map
(7) |
(8) |
(Eq. 7, Eq. 8). Here the matrix contains information about the weights between channels, e.g. represents the impact of channel on channel. In addition, we multiply the matrix between the transpose of and
(9) |
(10) |
, and reshape the result to . Then we multiply the result by a scale parameter and perform an element-wise sum operation with to obtain the final output (Eq. 9 and Eq. 10). After the CA module, all feature maps are summed and convolved by kernel to generate a weighted feature map, and we use the weighted feature map to re-weight by element-wise multiplying () it with the input (Eq. 11 and Eq. 12).
(11) |
(12) |
At last, we obtain the . According to the above procedure, we can input the extracted four scales feature maps , …, , into MAA module to get , …, , , respectively, as shown in Fig. 2.
Instead, the ASPP employs a series of dilated convolution layers with different rates to obtain larger receptive fields, however, dilated convolutions may lose local detail information. In contrast, the MAA module utilizes asymmetric convolutions with large kernels, which allows capturing multi-scale features without sacrificing too much detail. Furthermore, before the 1x1 convolution compresses the channels in the MAA module, a channel attention mechanism is adopted, enhancing the module’s focus on the channel dimension and better preserving important features.
Method | Indicators | Tissue Layers | |||||||
---|---|---|---|---|---|---|---|---|---|
NFL | GCL | INL/OPL | ONL | OS | RPE | ||||
ReLayNet[11] | 93.6 | 85.6 | 85.5 | 93.5 | 87.9 | 94.1 | 97.3 | 82.1 | |
93.3 | 82.1 | 86.0 | 97.2 | 87.5 | 91.1 | ||||
OS_MGU[30] | 94.8 | 86.8 | 86.7 | 94.1 | 88.6 | 94.5 | 97.7 | 83.6 | |
93.5 | 82.8 | 87.4 | 97.8 | 88.1 | 91.2 | ||||
DFANet[31] | 91.1 | 80.0 | 80.8 | 88.3 | 73.5 | 88.8 | 95.1 | 72.6 | |
91.3 | 77.0 | 82.9 | 93.3 | 74.4 | 86.3 | ||||
BiSeNet[32] | 94.9 | 87.3 | 88.3 | 93.9 | 84.8 | 94.5 | 97.8 | 83.1 | |
95.4 | 83.9 | 89.7 | 96.6 | 84.9 | 91.2 | ||||
Attention_Unet[12] | 94.6 | 87.4 | 87.6 | 93.6 | 87.1 | 93.7 | 97.6 | 83.1 | |
94.6 | 83.4 | 87.6 | 97.6 | 87.5 | 91.2 | ||||
EMV-Net[22] | 94.6 | 85.3 | 85.5 | 93.7 | 87.9 | 94.5 | 97.6 | 82.5 | |
93.5 | 81.9 | 86.2 | 97.1 | 90.1 | 91.7 | ||||
U-net [10] | 92.6 | 82.5 | 83.4 | 91.5 | 82.6 | 92.3 | 96.8 | 78.1 | |
90.3 | 80.8 | 84.8 | 96.6 | 80.8 | 90.0 | ||||
SegFormer-B0 [20] | 94.5 | 83.4 | 84.1 | 93.4 | 88.2 | 94.2 | 97.4 | 81.6 | |
92.8 | 79.0 | 86.9 | 97.6 | 86.3 | 91.3 | ||||
TransUnet[18] | 94.4 | 86.9 | 87.5 | 94.1 | 87.5 | 94.7 | 97.7 | 83.5 | |
93.4 | 83.2 | 88.0 | 97.0 | 87.2 | 91.6 | ||||
LightReSeg (Our) | 94.3 | 87.5 | 89.1 | 94.7 | 89.1 | 94.5 | 97.8 | 84.6 | |
94.3 | 84.5 | 89.4 | 96.9 | 87.8 | 91.6 |
III-D Light-weight Designs
The basic principle of our design of LightReSeg is to improve the overall performance of the model with as few parameters as possible, so we can meet the computational efficiency of deployments for real-time clinical applications. Thus, we make some light-weight designs for part of the structure of the model. For encoder extraction of multi-scale feature maps, we use depthwise separable convolutions (DS-Conv) for our down-sampling[14]. DS-Conv considers the channels and spatial regions of the feature map separately, takes different convolution kernels for different channels, and then uses point convolution to aggregate the channel information, achieving a richer feature representation with fewer parameters to learn. We find that using standard convolution for down-sampling is better than pooling, but convolution increases the number of parameters, so we use DS-Conv to ensure the segmentation performance while decreasing the burden of the number of parameters. In addition, as mentioned in Sec. III-C, multi-scale asymmetric convolution is used in our MAA module. Initially, our idea is to perform multi-scale extraction of feature information like ASPP [28], but we find that using standard convolution would cause a greater burden on the number of parameters, since the MAA module is supposed to perform extraction of the scale feature maps, which would certainly further increase the number of parameters. In order to reduce the number of parameters in the MAA module, we use asymmetric convolution instead of standard convolution, shown in Fig. 3, and the overall performance of LightReSeg is nearly the same. In particular, the Transformer block has a significant proportion of the number of parameters in LightReSeg. We have optimized the internal parameters of the Transformer block structure, such as the setting of heads=8 for multiple heads in MSA and dim_head=64 for hidden linear layers in MSA[13], etc. All these designs to minimize the number of parameters are made under the precondition of ensuring the performance of LightReSeg.
IV EXPERIMENT
In this section, we compare our LightReSeg against the state-of-the-art methods for retinal layer segmentation on i) two publicly available OCT datasets with unhealthy eyes, where the Glaucoma dataset is for the Glaucoma disease and the DME dataset is for Diabetic Macular Edema (DME), and ii) a new dataset, named as Vis-105H, that we create on top of raw images collected from [33] which correspond to only healthy human eyes with visible light OCT. Moreover, we present a thorough ablation study with our Vis-105H dataset to validate the effectiveness of the proposed MAA and the introduction of the Transformer block.
IV-A Datasets
Vis-105H dataset
We create the Vis-105H dataset for the method evaluation which contains images captured using a custom visible light OCT device that takes optic disc cubes from each subject [34, 35, 36], as shown in Fig. 4(a). The raw images are originally collected and presented in [33], and they correspond to 21 eyes from 14 healthy subjects, all of whom with written informed consent prior to the participation. The inclusion criteria include i) age years, ii) best corrected visual acuity better than 20/40, iii) no previous intraocular surgery of any kind, and iv) no known retinal disease, resulting in 21 scanning cubes with 5376 B-scan images. Two professional ophthalmologists further select 5 images with high imaging quality from every 256 B-scans, as shown in Fig. 4(b), with the selection criteria including i) low noise, ii) clear retinal layer boundaries, and iii) suitability for ophthalmologists to use for diagnosis. With the supervision of medical ophthalmology professionals, we annotate these 105 pre-processed retinal Visible-light OCT images of healthy human eyes for semantic segmentation, and complete the Vis-105H dataset for 7-class semantic segmentation including the background, as shown in Fig. 4(c). Specifically, we exploit the graphics software Inkscape to annotate the six layers, nerve fiber layer (NFL), ganglion cell layer (GCL) and inner plexiform layer (IPL), inner nuclear layer (INL) and out plexiform layer (OPL), outer nuclear layer (ONL), outer segment (OS), and retinal pigment epithelium (RPE), as red, yellow, green, light blue, dark blue, and pink, respectively, and to annotate the remaining regions as background black. Each annotated image is further examined by a professional ophthalmologist and exported in PNG format with a size of 660300 pixels. Fig. 4(d) shows the average pixel percentages of all categories on the OCT image except for the background, from which we can observe that the most dominant two classes are the NFL and ONL layers while the OS and GCL layers are the least dominant classes. The Vis-105H dataset is divided into a training set, a validation set, and a test set consisting of 75, 15, and 15 images, respectively. The data for the Vis-105 dataset is conducted at BMC, and both BU and BMC hold ownership rights to the data.
Glaucoma dataset
The dataset is collected from 61 different subjects [30], where 12 radial OCT B scans of each subject are collected using DRI OCT-1 Atlantis at the Ophthalmology Department of Shanghai General Hospital. All images are scanned at 20.48 mm × 7.94 mm field of view in the optic nerve head region. Under the supervision of a glaucoma specialist, two ophthalmologists manually annotate these images into the optic disc and nine retinal layers. The image size is 1024 × 992 and the dataset is divided into training, validation, and test sets by 148, 48, and 48 images, respectively.
DME dataset
The dataset is collected by Chiu et al. using the Duke Enterprise Data Unified Content Explorer search engine to retrospectively identify DME subjects within the Duke Eye Center[37]. It consists of 110 OCT B-scans obtained from 10 patients with DME with a size of 496768 pixels, each B-scan is annotated with 9 layers. The dataset is divided by 88, 11, 11 into training, validation, and test set, respectively. All subjects are scanned with the macula at the center.
Method | Indicators | Tissue Layers | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
NFL | GCL/IPL | INL | OPL | ONL/ISM | ISE | OS/RPE | Fluid | ||||
ReLayNet[11] | 82.0 | 93.4 | 79.4 | 77.2 | 86.8 | 86.6 | 86.5 | 53.8 | 95.5 | 69.6 | |
79.5 | 93.3 | 76.2 | 76.7 | 87.7 | 89.7 | 87.2 | 52.0 | ||||
OS_MGU[30] | 82.6 | 93.6 | 79.5 | 78.4 | 87.2 | 87.1 | 85.6 | 57.7 | 95.6 | 69.8 | |
80.7 | 93.7 | 77.0 | 77.2 | 88.3 | 89.5 | 84.6 | 52.1 | ||||
DFANet[31] | 77.4 | 90.3 | 73.6 | 72.6 | 85.8 | 85.7 | 85.5 | 49.8 | 94.7 | 64.8 | |
73.7 | 91.4 | 70.4 | 70.8 | 84.1 | 89.4 | 85.9 | 47.3 | ||||
BiSeNet[32] | 80.9 | 92.7 | 77.9 | 75.3 | 86.3 | 86.1 | 86.6 | 56.7 | 95.4 | 68.3 | |
76.7 | 93.0 | 76.8 | 72.2 | 86.9 | 87.5 | 89.8 | 51.0 | ||||
Attention_Unet[12] | 80.5 | 91.5 | 77.6 | 75.9 | 86.8 | 86.6 | 86.4 | 55.6 | 95.3 | 68.0 | |
79.4 | 92.4 | 74.4 | 74.0 | 86.8 | 88.9 | 86.6 | 49.9 | ||||
EMV-Net[22] | 80.4 | 92.5 | 78.5 | 74.8 | 85.6 | 86.3 | 86.3 | 59.9 | 95.3 | 68.4 | |
74.9 | 91.6 | 81.8 | 71.3 | 81.2 | 89.4 | 87.6 | 60.2 | ||||
U-net [10] | 82.2 | 92.8 | 79.1 | 78.2 | 86.7 | 87.4 | 86.6 | 56.2 | 95.5 | 69.4 | |
81.4 | 92.9 | 76.2 | 76.5 | 87.6 | 90.7 | 86.7 | 52.2 | ||||
SegFormer-B0 [20] | 81.7 | 93.7 | 79.1 | 77.0 | 86.6 | 86.6 | 85.7 | 57.6 | 95.6 | 69.2 | |
77.5 | 95.0 | 74.9 | 77.0 | 89.8 | 89.9 | 87.8 | 49.1 | ||||
TransUnet[18] | 81.4 | 93.1 | 79.8 | 77.4 | 87.4 | 87.2 | 86.6 | 60.1 | 95.7 | 70.0 | |
80.0 | 93.0 | 77.7 | 77.6 | 88.2 | 89.5 | 86.7 | 53.9 | ||||
LightReSeg (Our) | 82.4 | 93.7 | 79.8 | 78.8 | 87.9 | 87.3 | 86.0 | 59.7 | 95.7 | 70.5 | |
79.7 | 93.5 | 77.1 | 78.1 | 87.8 | 89.6 | 85.6 | 57.8 |
Method | Indicators | Tissue Layers | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NFL | GCL | IPL | INL | OPL | ONL | IS/OS | RPE | Choroid | OD | ||||
ReLayNet[11] | 79.4 | 64.9 | 70.8 | 76.4 | 81.1 | 90.5 | 86.0 | 93.1 | 87.9 | 77.8 | 94.2 | 67.0 | |
77.2 | 60.7 | 70.6 | 77.0 | 80.5 | 90.0 | 84.9 | 85.2 | 88.6 | 77.4 | ||||
OS-MGU[30] | 80.7 | 62.0 | 69.6 | 74.2 | 78.4 | 89.6 | 84.7 | 83.0 | 88.5 | 83.7 | 94.8 | 66.7 | |
82.8 | 56.1 | 68.7 | 74.4 | 75.7 | 88.3 | 85.0 | 85.7 | 88.7 | 86.8 | ||||
DFANet[31] | 79.8 | 60.5 | 68.2 | 74.3 | 77.4 | 89.4 | 85.0 | 81.3 | 88.2 | 83.9 | 94.7 | 65.8 | |
80.0 | 55.9 | 66.1 | 71.9 | 73.1 | 91.7 | 85.7 | 80.8 | 89.6 | 88.6 | ||||
BiSeNet[32] | 78.0 | 61.5 | 67.4 | 70.1 | 77.1 | 88.5 | 82.4 | 78.3 | 86.9 | 84.3 | 94.5 | 64.0 | |
76.4 | 57.0 | 65.6 | 66.2 | 77.4 | 89.5 | 82.3 | 81.6 | 90.1 | 89.3 | ||||
Attention_Unet[12] | 79.5 | 60.1 | 64.7 | 72.8 | 78.8 | 90.6 | 86.3 | 82.8 | 88.8 | 83.2 | 94.7 | 66.0 | |
84.2 | 57.9 | 59.4 | 71.0 | 75.0 | 91.9 | 85.6 | 81.8 | 89.1 | 84.8 | ||||
EMV-Net[22] | 81.2 | 65.3 | 69.5 | 72.1 | 79.1 | 90.0 | 83.9 | 81.5 | 88.5 | 83.4 | 94.8 | 66.6 | |
83.7 | 59.3 | 67.8 | 65.8 | 74.8 | 88.6 | 81.5 | 83.4 | 89.9 | 86.8 | ||||
U-net [10] | 81.9 | 63.7 | 69.4 | 73.5 | 77.5 | 89.7 | 85.4 | 82.3 | 87.8 | 83.3 | 94.8 | 66.6 | |
83.5 | 58.1 | 68.3 | 73.7 | 74.6 | 91.7 | 84.2 | 83.4 | 91.0 | 86.5 | ||||
SegFormer-B0 [20] | 82.1 | 64.3 | 69.4 | 76.4 | 76.6 | 89.9 | 84.4 | 81.2 | 88.6 | 84.2 | 94.8 | 66.9 | |
86.7 | 58.5 | 63.9 | 81.7 | 69.7 | 92.6 | 82.8 | 79.8 | 87.7 | 90.7 | ||||
TransUnet[18] | 80.6 | 60.0 | 69.4 | 76.1 | 80.2 | 90.8 | 85.8 | 82.6 | 88.5 | 82.4 | 94.7 | 67.0 | |
84.7 | 53.3 | 69.0 | 77.6 | 78.5 | 93.6 | 83.0 | 80.3 | 86.9 | 84.3 | ||||
LightReSeg (Our) | 80.9 | 64.3 | 70.5 | 77.2 | 80.4 | 90.1 | 85.4 | 82.7 | 88.4 | 83.3 | 94.9 | 67.8 | |
83.5 | 61.1 | 70.8 | 78.8 | 78.7 | 90.2 | 85.0 | 85.3 | 90.5 | 83.5 |
IV-B Performance Metrics
We evaluate the segmentation performance quantitatively using the Dice similarity coefficient (DSC), mean Intersection over Union (mIoU), pixel accuracy (PA), and mean pixel accuracy (mPA). They can be computed as:
(13) |
(14) |
(15) |
and
(16) |
where , , , and represent True Positives, False Positives, True Negatives, and False Negatives respectively. represents the number of categories. In addition, we count the number of parameters for compared approaches to represent their computation complexity.
IV-C Implementation Details
LightReSeg is implemented based on the PyTorch and trained with the Adam optimizer with the cross-entropy loss. The initial learning rate is set to 0.001 which is then gradually halved every 40 epochs. Data augmentation is applied on all three datasets, including horizontal flipping with probability P=0.5, random center rotation within plus or minus 20 degrees with probability P=0.5, median and motion blur processing with P=0.5 probability, random Gaussian noise addition, as well as random brightness and contrast with P=0.5 probability. All experiments are conducted on an NVIDIA GeForce RTX 3090 Graphics card. The code of the proposed LightReSeg could be found at: https://github.com/Medical-Image-Analysis/LightReSeg.
IV-D Comparison
We compare LightReSeg with the state-of-the-art approaches including ReLaynet[11], EMV-Net[22], Attention_Unet[12], BiSeNet[32], DFANet[31], OS_MGU[30], U-net [10] and TransUnet[18]. We report the results on the above-mentioned three datasets.
IV-D1 Quantitative Analysis
Tab. I shows the comparison of our approach with state-of-the-art methods on the Vis-105H dataset. We can see that our approach LightReSeg scores the best performance in both and metrics, with improvement in terms of the compared to the second-best performing method OS_MGU [30]. In terms of the metric, our approach achieves the highest segmentation performance in the GCL layer, INL/OPL layer, ONL layer, and OS layer. Regarding the metric, our approach achieves the first place in the GCL layer, and the second place in the INL/OPL layer and RPE layer, which also shows the outstanding performance of our proposed approach. The segmentation performance of ReLayNet, the lightest algorithm in the field of retinal layer segmentation, is lower than our proposed LightReSeg in terms of the metric. We further perform a statistical significance test, using the Wilcoxon rank sum test, to compare the Dice Score performance of different methods on each layer. When comparing our approach with ReLayNet, we observe a of 0.031250 (p0.05), indicating a statistically significant difference. The poor performance of ReLayNet might be due to its rather simple feature fusion strategy for combining features from encoder and decoder, i.e. skip connections, leading to limited semantic information at the fusion. Instead, LightReSeg further introduces the MAA module for the feature fusion at multiple scales. Moreover, we add Transformer layers at the bottom of the encoder for efficient global reasoning which further improves the segmentation accuracy. SegFormer is a new light-weight model in the field of semantic segmentation and has a good performance in many tasks [20], but as can be seen from Tab. I, SegFormer-B0 only ranks 7th in terms of the and metrics, while in terms of the metric, its segmentation performance of all layers is much lower than ours. In terms of the metric, only the ONL layer is higher than our by 0.7 percentage points, and the segmentation performance in the rest of the layers is much lower than our approach. We analyze that the poor segmentation performance of SegFormer-B0 on this task may be mainly due to two factors: first, its excessive pursuit of lightweight results in limited ability to extract detailed features; second, the up-sampling part of SegFormer-B0’s decoder is too rough, resulting in the lack of fine detail recovery.

We evaluate our approach on the DME dataset, as shown in Tab. II. We achieve the best performance in terms of both and . In terms of we achieve first place in 4 layers (i.e. GCL/IPL, INL, OPL, and ONL/ISM) and second place in another 2 layers (i.e. NFL and ISE), while in terms of , we are on par with the state-of-the-art performance. Among other approaches, such as TransUnet, OS_MGU, and ReLayNet, although they are more advanced in the metric, they are 0.5, 0.7, and 0.9 percentage points lower than the first place in the metric, respectively, which also indicates that our model also has a strong comprehensive performance on the DME dataset. We further perform the statistical significance test by using the Wilcoxon rank sum test. For example, when comparing our method with TransUnet on the same domain, we observe a of 0.017629 (p0.05), indicating a statistically significant difference. Similar statistically significant differences are observed, with of 0.023437 (p0.05), 0.039062 (p0.05), and 0.017755 (p0.05) respectively when comparing our method with OS_MGU, EMV-Net, and SegFormer-B0. We further evaluate our approach on the Glaucoma dataset, as shown in Tab. III. Our LightReSeg achieves the best results on both the overall metrics and . Additionally, our approach also scores better on most of the layers in terms of both and .
Fig. 5 (a) and (b) show the performance of the ten mainstream approaches on the three datasets measured by the mIoU and mPA metrics, respectively. Our approach LightReSeg achieve the best performance on all three datasets in terms of and . TransUnet performs is the second best-performing method with a small margin lower than ours on all three datasets, however, its number of parameters is 32 times of the one of our approach. The lightest model ReLayNet is inferior to our method in terms of performance by , , and on the Vis-105H, DME, and Glaucoma dataset, respectively. Therefore, our approach achieves state-of-the-art performance with a relatively light-weight model.
In addition, we find that regardless of the method, certain layers (NFL, ONL, RPE) have significantly higher accuracy compared to other layers, as shown in Tab. I. We believe that class imbalance is the most likely cause, as the number of samples in a certain class is much larger than in other classes, resulting in the model learning better for that class during training and exhibiting higher accuracy in evaluation. From Fig. 4(d), we can see that the average pixel proportion of NFL, ONL, and RPE layers is relatively high, and Tab. I also shows that these three layers have higher accuracy compared to other layers. Also, there are differences in shape, color, texture, and other aspects among different layers, which may make certain categories of targets easier to segment while others are more difficult. In summary, we believe that the best way to solve this problem is to first address class imbalance.


IV-D2 Qualitative Analysis
Fig. 6 shows the segmentation prediction maps of several mainstream segmentation methods in a retinal B-scan image that contains blood flow information disturbance. As shown in Fig. 6(j), our approach achieves much better segmentation performance. The prediction graphs of all the approaches in Fig. 6(c), (d), (f), (h) all showed false positive errors, specifically, the region above the NFL layer or below the RPE layer of the retina is incorrectly identified as the retinal layer and segmented, resulting in the wrong segmentation target. The TransUnet in Fig. 6(e) performs well, but two locations on the boundary between the NFL and GCL layers are inaccurately segmented, as marked by the two locations in Fig. 6(e). The same occurs in EMV-Net in one place, as shown in Fig. 6(g). In comparison, our approach is not only without false positive errors, but also segmentation boundaries are more accurate.
Fig. 7 shows the comparative prediction plots of several mainstream segmentation approaches in a single B-scan image over the retinal optic nerve head region. As shown in Fig. 7(j), our approach shows a better segmentation performance. The black dashed region in Fig. 7(d), where the NFL and GCL layer boundaries appear distinctly jagged, and the black dashed region in Fig. 7(f), where a distinct honeycomb shape appears, are signs of unsmooth and inaccurate layer boundary prediction, and the same occurs in Fig. 7(c), (e), and (h). In addition, the problem of segmentation category error occurs, such as the red dashed area in Fig. 7(d), where the segmentation results in incorrectly segmenting part of the OPL layer into ONL layers, and this segmentation category error occurs at least twice in Fig. 7, (e), (f), and (g). In contrast, our approach not only has no problem with segmentation category error but also the segmented boundary is smoother than other approaches.
IV-E Ablation study
IV-E1 Performance of Different Modules
We conduct ablation experiments on the Vis-105H dataset with the main purpose of verifying the effect of the MAA module and the contribution of the Transformer block on the overall method segmentation performance. The ablated methods are: Base is the U-shaped backbone; Base_MAA and Base_trans3 are MAA and 3 transformer layers added to Base, respectively; Base_MAA_trans3 and Base_MAA_trans6 are based on Base_MAA with 3 and 6 transformer layers respectively. Note that Base_MAA_trans3 represents the final version of our proposed LightReSeg. As shown in Tab. IV, the Base_MAA approach improves over the Base approach, the and by and , respectively. All six layers except for the OS layer are also significantly improved in the . We also observe that Base_MAA_trans3 improves over Base_trans3 in terms of both and . Fig. 8(c) and (d) show that the segmentation performance of the model is greatly improved by adding the MAA module to the model, and the results of mis-segmentation are significantly reduced in the background region below the RPE layer.
Regarding the Transformer block, as shown in Tab. IV, Base_trans3 improves and over Base in terms of and , respectively, and the segmentation ability of each layer is improved in terms of . In addition, we also find that increasing the number of Transformer layers can further improve the segmentation performance within a certain range. For example, Base_MAA_trans6 adds 3 more Transformer layers than Base_MAA_trans3, which improves and by and , respectively, this comes at a cost of more of parameters. Fig. 8(c) and (e) show qualitatively that adding three sequential Transformer blocks to the model contributes to a big improvement in the segmentation performance of the model, and there is no segmentation error in the background region below the RPE layer.
Method | (%) Results of Tissue Layers | (%) | (%) | Params (M) | |||||
---|---|---|---|---|---|---|---|---|---|
NFL | GCL | INL/OPL | ONL | OS | RPE | ||||
Base | 94.0 | 84.6 | 84.8 | 93.2 | 87.9 | 92.6 | 97.2 | 81.3 | 1.61 |
Base_MAA | 94.4 | 86.9 | 87.7 | 94.0 | 87.4 | 94.0 | 97.6 | 83.2 | 1.72 |
Base_trans3 | 94.7 | 86.9 | 87.3 | 94.1 | 89.0 | 94.0 | 97.7 | 83.7 | 3.20 |
Base_MAA_trans3(Our) | 94.3 | 87.5 | 89.1 | 94.7 | 89.1 | 94.5 | 97.8 | 84.6 | 3.30 |
Base_MAA_trans6 | 95.0 | 87.6 | 89.0 | 94.8 | 89.7 | 95.1 | 97.9 | 85.2 | 4.68 |

To further interpret the role of different modules, we use a modified version of Grad-CAM[38] to visualize the feature activations in different layers of the model (see Fig. 9). Specifically, in accordance with the model settings in Tab. IV, we extract the feature activation heat maps for different retinal layers at different model structures and at different positions of the model, respectively. By comparing Fig. 9(e) and (i), with the addition of the MAA module, we can see that the heat map of the corresponding NFL layer in Fig. 9(i) is further enhanced compared to Fig. 9(e). The same trend is also observed in Fig. 9(h) and (l). It is further found that with the addition of the MAA module to the Base_trans3 model structure, the feature region corresponding to the segmentation category is also further enhanced, as shown in Fig. 9(p) and (t). Adding the Transformer blocks to the model also causes changes in the heat map, as in Fig. 9(f) and (n). The heat map of the corresponding RPE layer region is activated more comprehensively with the addition of the Transformer blocks, and the OS and background layers of the adjacent regions also show different degrees of activation.

When analyzing the activation maps, we observe that not only the regions corresponding to the predicted layers prominently highlighted, but also the feature weights of the adjacent layers are also significantly amplified. To interpret this behavior, we conduct further analysis and find that the boundaries of the retinal layers exhibit a high degree of correlation with the adjacent layers. As in Fig. 9(e), the focus is mainly on the NFL layer, but adjacent layers (e.g., background and GCL) also show various degrees of activation. Similarly, in Fig. 9(r), the RPE layer of the retinal layer is the main focus, while information from the OS and background layers at the upper and lower boundaries of the retinal layer is also utilized. This phenomenon suggests that the model uses information from adjacent layers to enhance its predictions. In addition, we also export the feature activation maps for different deep network layers in the prediction of the RPE layer, and we find that the heat map of the deep network focuses on more detailed and accurate regions compared to the heat map of the superficial network. As in Fig. 9(o), the regions where the heat map is focused on activation are highly coincident with the RPE layer, while the regions activated by the heat map in Fig. 9(n) are much coarser, which is similarly shown in Fig. 9(r) and (s). This observation suggests that the deeper extract image features of the model network contain more accurate and insightful semantic information. Overall, the visualization of deep feature activation provides valuable insights into the model’s decision process, and the model can utilize multiple layers of information to obtain more robust segmentation results.
IV-E2 Light-weight Setup
To evaluate the potential impact of DS-Conv on feature extraction capabilities, we replace all DS-Conv in the encoder section with standard convolutions and conduct comparative analyses[39], as shown in Tab. V. The results indicate that using standard convolutions instead of DS-Conv does not enhance feature extraction in our optimized model and leads to an 11% increase in the parameter count. This demonstrates that a simple increase in the model’s parameter count does not always result in performance.
Method Settings | (%) | (%) | Params(M) |
---|---|---|---|
Encoder with DS-Conv | 97.8 | 84.6 | 3.30 |
Encoder with Conv | 97.8 | 84.3 | 3.69 |
For assessing the impact of model light-weight on performance, we modify the number of channels in the encoder’s multi-scale features, both halving and doubling them, the specific results are shown in Tab. VI. We find that, based on our model, when the channel of the multi-scale features extracted by the encoder is reduced to 0.5 and 0.25 times, the model’s parameter count decreased, but its segmentation performance significantly deteriorated. When we increase the channel count to 2 and 4 times, there is no significant change in the model’s accuracy, yet the parameter counts are 3 and 11 times the original size, respectively. From this, we can also conclude that our lightweight design maintains accuracy while using the fewest possible parameters.
Method Settings | mPA(%) | mIoU(%) | Params(M) |
---|---|---|---|
0.25x | 96.9 | 79.2 | 0.48 |
0.5x | 96.9 | 79.1 | 1.18 |
1x | 97.8 | 84.6 | 3.30 |
2x | 97.8 | 84.3 | 9.87 |
4x | 97.7 | 84.1 | 33.48 |
IV-F Limitation
Although LightReSeg boasts only 3.3M trainable parameters and delivers the optimal retinal layer segmentation outcomes, in real-world applications, model parameter size is not the sole determinant of algorithm efficacy. Inference speed emerges as a crucial metric for assessing algorithmic effectiveness. As evidenced by Tab. VII, our proposed method exhibits inference speeds of 0.11, 0.27, and 0.07 on the three datasets, respectively. While these figures already signify remarkable efficiency, there remains scope for further enhancement. For example, although the segmentation accuracy of ReLayNet is far inferior to our approach, it must be admitted that its inference efficiency is slightly higher than ours.
Method |
|
|
|
|
||||
---|---|---|---|---|---|---|---|---|
ReLayNet[11] | 0.7 | 0.08 | 0.20 | 0.06 | ||||
EMV-Net[22] | 1.9 | 0.15 | 0.31 | 0.13 | ||||
OS_MGU[30] | 2.0 | 0.12 | 0.30 | 0.07 | ||||
DFANet[31] | 2.1 | 0.12 | 0.31 | 0.08 | ||||
SegFormer-B0 [20] | 3.7 | 0.12 | 0.28 | 0.07 | ||||
BiSeNet[32] | 13.1 | 0.08 | 0.20 | 0.05 | ||||
U-net [10] | 34.5 | 0.10 | 0.20 | 0.08 | ||||
Attention_Unet[12] | 34.8 | 0.10 | 0.22 | 0.09 | ||||
TransUnet[18] | 105.7 | 0.14 | 0.32 | 0.11 | ||||
LightReSeg (Our) | 3.3 | 0.11 | 0.27 | 0.07 |
Moreover, in the context of practical clinical applications, constraints related to dataset limitations preclude the inclusion of all device data types. Imagery obtained through disparate devices may exhibit domain discrepancies, inevitably resulting in false positive errors in the segmentation output for unfamiliar new devices. Consequently, apart from enlarging the dataset, enhancing the model’s domain generalization competency is a subject worthy of further investigation. The supplementary segmentation result evaluation system will also augment ophthalmologists’ confidence in the segmentation outcomes.
V CONCLUSION
In this paper, we propose a novel light-weight method LightReSeg for retinal layer segmentation. Our method introduces a Transformer-based block in the encoder part to enable global reasoning for reducing errors in the background region and an attention mechanism, named MAA module, to best exploit rich semantic information for the multi-scale feature fusion to improve the segmentation accuracy. The extensive evaluation shows that our approach achieves the best segmentation performance on both our collected Vis-105H dataset and two other public ones, i.e. DME and Glaucoma, indicating that our model has good reliability in the face of noise and uncertainty in the data. To improve the efficiency, our method also incorporates light-weight designs. The ablation experiments detailed in Section IV-E2 demonstrate that our lightweight design is precisely adequate, maintaining the highest level of accuracy while significantly reducing the number of parameters compared to other high-performing methods. This further highlights the practicality of our proposed approach, particularly evident in the real-time preview functionality of small-field-of-view OCTA imaging. In the future, we will collect more clinical data to further improve our approach and continuously contribute to the efficient performance of OCT devices.
References
- [1] F. K. Horn, C. Y. Mardin, R. Laemmer, D. Baleanu, A. M. Juenemann, F. E. Kruse, and R. P. Tornow, “Correlation between local glaucomatous visual field defects and loss of nerve fiber layer thickness measured with polarimetry and spectral domain oct,” Investigative ophthalmology & visual science, vol. 50, no. 5, pp. 1971–1977, 2009.
- [2] F. Pollet-Villard, C. Chiquet, J.-P. Romanet, C. Noel, and F. Aptel, “Structure-function relationships with spectral-domain optical coherence tomography retinal nerve fiber layer and optic nerve head measurements,” Investigative ophthalmology & visual science, vol. 55, no. 5, pp. 2953–2962, 2014.
- [3] A. Ajaz, H. Kumar, and D. Kumar, “A review of methods for automatic detection of macular edema,” Biomedical Signal Processing and Control, vol. 69, p. 102858, 2021.
- [4] C. Brandl, C. Brücklmayer, F. Günther, M. E. Zimmermann, H. Küchenhoff, H. Helbig, B. H. Weber, I. M. Heid, and K. J. Stark, “Retinal layer thicknesses in early age-related macular degeneration: results from the german augur study,” Investigative ophthalmology & visual science, vol. 60, no. 5, pp. 1581–1594, 2019.
- [5] J. A. van de Kreeke, H.-T. Nguyen, J. den Haan, E. Konijnenberg, J. Tomassen, A. den Braber, M. Ten Kate, L. Collij, M. Yaqub, B. van Berckel et al., “Retinal layer thickness in preclinical alzheimer’s disease,” Acta ophthalmologica, vol. 97, no. 8, pp. 798–804, 2019.
- [6] D. Huang, E. A. Swanson, C. P. Lin, J. S. Schuman, W. G. Stinson, W. Chang, M. R. Hee, T. Flotte, K. Gregory, and C. A. Puliafito, “Optical coherence tomography,” Science, vol. 254, no. 5035, pp. 1178–1181, 1991.
- [7] X. Shu, L. J. Beckmann, and H. F. Zhang, “Visible-light optical coherence tomography: a review,” Journal of biomedical optics, vol. 22, no. 12, p. 121707, 2017.
- [8] W. Song, W. Shao, and J. Yi, “Wide-field and micron-resolution visible light optical coherence tomography in human retina by a linear-k spectrometer,” in Bio-Optics: Design and Application. Optica Publishing Group, 2021, pp. DM2A–4.
- [9] W. Song, W. Shao, W. Yi, and J. Yi, “Linear k-domain wide-field and micron-resolution visible light human retinal optical coherence tomography,” in Ophthalmic Technologies XXXII. SPIE, 2022, p. PC1194102.
- [10] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention. Springer, 2015, pp. 234–241.
- [11] A. G. Roy, S. Conjeti, S. P. K. Karri, D. Sheet, A. Katouzian, C. Wachinger, and N. Navab, “Relaynet: retinal layer and fluid segmentation of macular optical coherence tomography using fully convolutional networks,” Biomedical optics express, vol. 8, no. 8, pp. 3627–3642, 2017.
- [12] O. Oktay, J. Schlemper, L. L. Folgoc, M. Lee, M. Heinrich, K. Misawa, K. Mori, S. McDonagh, N. Y. Hammerla, B. Kainz et al., “Attention u-net: Learning where to look for the pancreas,” arXiv preprint arXiv:1804.03999, 2018.
- [13] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- [14] F. Chollet, “Xception: Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1251–1258.
- [15] X. Ding, Y. Guo, G. Ding, and J. Han, “Acnet: Strengthening the kernel skeletons for powerful cnn via asymmetric convolution blocks,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 1911–1920.
- [16] S. Wang, L. Yu, X. Yang, C.-W. Fu, and P.-A. Heng, “Patch-based output space adversarial learning for joint optic disc and cup segmentation,” IEEE transactions on medical imaging, vol. 38, no. 11, pp. 2485–2495, 2019.
- [17] J. Li, G. Gao, L. Yang, G. Bian, and Y. Liu, “Dpf-net: A dual-path progressive fusion network for retinal vessel segmentation,” IEEE Transactions on Instrumentation and Measurement, 2023.
- [18] J. Chen, Y. Lu, Q. Yu, X. Luo, E. Adeli, Y. Wang, L. Lu, A. L. Yuille, and Y. Zhou, “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021.
- [19] M. Astaraki, Ö. Smedby, and C. Wang, “Prior-aware autoencoders for lung pathology segmentation,” Medical Image Analysis, p. 102491, 2022.
- [20] E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, “Segformer: Simple and efficient design for semantic segmentation with transformers,” Advances in neural information processing systems, vol. 34, pp. 12 077–12 090, 2021.
- [21] B. Chen, T. Niu, W. Yu, R. Zhang, Z. Wang, and B. Li, “A-net: An a-shape lightweight neural network for real-time surface defect segmentation,” IEEE Transactions on Instrumentation and Measurement, 2023.
- [22] X. He, Y. Wang, F. Poiesi, W. Song, Q. Xu, Z. Feng, and Y. Wan, “Exploiting multi-granularity visual features for retinal layer segmentation in human eyes,” Frontiers in Bioengineering and Biotechnology, vol. 11, p. 1191803, 2023.
- [23] B. Wang, W. Wei, S. Qiu, S. Wang, D. Li, and H. He, “Boundary aware u-net for retinal layers segmentation in optical coherence tomography images,” IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 8, pp. 3029–3040, 2021.
- [24] X. Liu, J. Cao, S. Wang, Y. Zhang, and M. Wang, “Confidence-guided topology-preserving layer segmentation for optical coherence tomography images with focus-column module,” IEEE Transactions on Instrumentation and Measurement, vol. 70, pp. 1–12, 2020.
- [25] K. Hu, D. Liu, Z. Chen, X. Li, Y. Zhang, and X. Gao, “Embedded residual recurrent network and graph search for the segmentation of retinal layer boundaries in optical coherence tomography,” IEEE Transactions on Instrumentation and Measurement, vol. 70, pp. 1–17, 2021.
- [26] M. Gende, V. Mallen, J. de Moura, B. Cordón, E. Garcia-Martin, C. I. Sánchez, J. Novo, and M. Ortega, “Automatic segmentation of retinal layers in multiple neurodegenerative disorder scenarios,” IEEE Journal of Biomedical and Health Informatics, 2023.
- [27] M.-H. Guo, C.-Z. Lu, Q. Hou, Z. Liu, M.-M. Cheng, and S.-M. Hu, “Segnext: Rethinking convolutional attention design for semantic segmentation,” arXiv preprint arXiv:2209.08575, 2022.
- [28] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 834–848, 2017.
- [29] J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, and H. Lu, “Dual attention network for scene segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 3146–3154.
- [30] J. Li, P. Jin, J. Zhu, H. Zou, X. Xu, M. Tang, M. Zhou, Y. Gan, J. He, Y. Ling, and Y. Su, “Multi-scale gcn-assisted two-stage network for joint segmentation of retinal layers and discs in peripapillary oct images,” Biomed. Opt. Express, vol. 12, no. 4, pp. 2204–2220, Apr 2021. [Online]. Available: http://opg.optica.org/boe/abstract.cfm?URI=boe-12-4-2204
- [31] H. Li, P. Xiong, H. Fan, and J. Sun, “Dfanet: Deep feature aggregation for real-time semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 9522–9531.
- [32] C. Yu, J. Wang, C. Peng, C. Gao, G. Yu, and N. Sang, “Bisenet: Bilateral segmentation network for real-time semantic segmentation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 325–341.
- [33] W. Song, S. Zhang, Y. M. Kim, N. Sadlak, M. G. Fiorello, M. Desai, and J. Yi, “Visible light optical coherence tomography of peripapillary retinal nerve fiber layer reflectivity in glaucoma,” Translational vision science & technology, vol. 11, no. 9, pp. 28–28, 2022.
- [34] W. Song, S. Zhang, N. Sadlak, M. G. Fiorello, M. Desai, and J. Yi, “Visible and near-infrared optical coherence tomography (vnoct) in normal, suspect, and glaucomatous eyes,” in Ophthalmic Technologies XXXI, vol. 11623. SPIE, 2021, p. 1162311.
- [35] J. Yi and W. Song, “Systems and methods for fiber-based visible and near infrared optical coherence tomography,” Aug. 4 2020, uS Patent 10,732,354.
- [36] W. Song, L. Zhou, S. Zhang, S. Ness, M. Desai, and J. Yi, “Fiber-based visible and near infrared optical coherence tomography (vnoct) enables quantitative elastic light scattering spectroscopy in human retina,” Biomedical Optics Express, vol. 9, no. 7, pp. 3464–3480, 2018.
- [37] S. J. Chiu, M. J. Allingham, P. S. Mettu, S. W. Cousins, J. A. Izatt, and S. Farsiu, “Kernel regression based segmentation of optical coherence tomography images with diabetic macular edema,” Biomedical optics express, vol. 6, no. 4, pp. 1172–1194, 2015.
- [38] K. Vinogradova, A. Dibrov, and G. Myers, “Towards interpretable semantic segmentation via gradient-weighted class activation mapping (student abstract),” in Proceedings of the AAAI conference on artificial intelligence, vol. 34, no. 10, 2020, pp. 13 943–13 944.
- [39] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.