Mitosis Detection from Partial Annotation
by Dataset Generation via Frame-Order Flipping
Abstract
Detection of mitosis events plays an important role in biomedical research. Deep-learning-based mitosis detection methods have achieved outstanding performance with a certain amount of labeled data. However, these methods require annotations for each imaging condition. Collecting labeled data involves time-consuming human labor. In this paper, we propose a mitosis detection method that can be trained with partially annotated sequences. The base idea is to generate a fully labeled dataset from the partial labels and train a mitosis detection model with the generated dataset. First, we generate an image pair not containing mitosis events by frame-order flipping. Then, we paste mitosis events to the image pair by alpha-blending pasting and generate a fully labeled dataset. We demonstrate the performance of our method on four datasets, and we confirm that our method outperforms other comparisons which use partially labeled sequences. Code is available at https://github.com/naivete5656/MDPAFOF.
Keywords:
Partial annotation Mitosis detection Fluorescent image.1 Introduction
Fluorescent microscopy is widely used to capture cell nuclei behavior. Mitosis detection is the task of detecting the moment of cell division from time-lapse images (the dotted circles in Fig. 2). Mitosis detection from fluorescent sequences is important in biological research, medical diagnosis, and drug development.
Conventionally tracking-based methods [21, 19, 1, 19] and tracking-free methods [3, 6, 5, 7] have been proposed for mitosis detection. Recently, deep-learning-based mitosis-detection methods have achieved outstanding performance [12, 18, 11, 9, 10, 8]. However, training deep-learning methods require a certain amount of annotation for each imaging condition, such as types of cells and microscopy and the density of cells. Collecting a sufficient number of labeled data covering the variability of cell type and cell density is time-consuming and labor-intensive.
Unlike cell detection and segmentation, which aims to recognize objects from a single image, mitosis detection aims to identify events from time series of images. Thus, it is necessary to observe differences between multiple frames to make mitosis events annotation. Comprehensively annotating mitosis events is time-consuming, and annotators may be missed mitosis events. Thus, we must carefully review the annotations to ensure that they are comprehensive.
Partial annotation has been used as a way to reduce the annotation costs of cell and object detection [2, 22, 15]. Fig. 2 shows an example of partially annotated frames. Some mitosis events are annotated (a red-dotted circle), and others are not (light-blue-dotted circles). The annotation costs are low because the annotator only needs to plot a few mitotic positions. In addition, this style of annotation allows for missing annotations. Therefore, it would be effective for mitosis detection.
Unlike supervised annotation, partial annotation can not treat unannotated areas as regions not containing mitosis events since the regions may contain mitosis events (Fig. 2). The regions naturally affect the training in the partial annotation setting. To avoid the effect of unlabeled objects in unlabeled regions, Qu et al. [15] proposed to use a Gaussian masked mean squared loss, which calculates the loss around the annotated regions. The loss function works in tasks in which foreground and background features have clearly different appearances, such as in cell detection. However, it does not work on mitosis detection since the appearance of several non-mitotic cells appears similar to mitosis cells; it produces many false positives.
![]() ![]() |
In this paper, we propose a cell-mitosis detection method for fluorescent time-lapse images by generating a fully labeled dataset from partially annotated sequences. We achieve mitosis detection training in a mitosis detection model with the generated dataset. To generate the fully labeled dataset, we should consider two problems: (1) no label indicating regions not containing mitosis cells and (2) few mitosis annotations.
We can easily generate the regions not containing mitotic cells by using one image twice. However, such regions do not contribute to identifying mitotic cells and non-mitotic cells since the data do not show natural cell motions. For the training to be effective, the regions not containing mitotic cells should show the natural movements of cells. To generate such regions, we propose frame-order flipping which simply flips the frame order of a consecutive frame pair. As shown in the white rectangles in Fig. 2, we can convert a mitosis event to a cell fusion by flipping operation. Hence, the flipped pair is the region not containing mitosis cells. Even though we flipped the frame order, the non-mitotic cells still have natural time-series motion, as shown in the yellow rectangles in Fig. 2.
In addition, we can make the most of a few partial annotations by using copy-and-paste-based techniques. Unlike regular copy-and-paste augmentation [4] for supervised augmentation of instance segmentations which have object mask annotations, we only have point-level annotations. Thus, we propose to use alpha-blending pasting techniques which naturally blend two images.
Experiments conducted on four types of fluorescent sequences demonstrate that the proposed method outperforms other methods which use partial labels.
Related work Some methods used partially labeled data to train model [15, 22, 2]. Qu [15] proposed a Gaussian masked mean squared loss, which calculates the loss around the annotated areas. To more accurately identify negative and positive samples, positive unlabeled learning has been used for object detection [22, 2]. These methods have used positive unlabeled learning on candidates detected by using partial annotation to identify whether the candidates are labeled objects or backgrounds. However, since the candidates detected by partial annotation include many false positives, the positive unlabeled learning does not work on mitosis detection. the appearance of the mitosis event and backgrounds in the mitosis detection task, it is difficult to estimate positive prior. These methods could not work on mitosis detection. The positive unlabeled learning requires a positive prior.

2 Method: mitosis detection with partial labels
Our method aims to detect coordinates and timing of mitosis events from fluorescent sequences. For training, we use time-lapse images and partial labels (a set of annotated mitosis cells). Here, denotes an image at frame , and is the total number of frames.
Our method generates a fully labeled dataset from time-lapse images and partial labels and then trains a mitosis detection model with the generated dataset. Here, is a generated image, and is a set of mitotic coordinates contained in . Since our method trains the network with partial labels, it can eliminate the costs of checking for missed annotations.
2.1 Labeled dataset generation
Fig. 3 shows an overview of our dataset generation. We randomly pick a pair of consecutive frames from time-lapse images . Since the pair may contain unannotated mitosis events, we forcibly convert the pair into a negative pair (i.e., a pair which does not contain mitosis events) by using frame-order flipping. Next, we paste mitosis events to a generated pair using alpha-blending pasting and obtain a generated pair . Since we know the pasted location, we can obtain the mitosis locations of the generated pair.

Negative pair generation with frame-order flipping: In this step, we generate a pair not containing mitotic cells by using a simple augmentation-based frame-order flipping. Fig. 3 shows an example of the pair images . The pair may contain mitosis events. If we assume that the pair does not contain mitotic cells, it affects the training of the mitosis detection model . To prevent the pair from containing mitosis events, we flip the frame order and treat the flipped pair as a pair of negative.
Since mitosis is the event that a cell divides into two daughter cells, the mitosis event is transformed into an event in which two cells fuse into one by flipping the order (Fig. 2). The flipped event can treat as a non-mitotic event. Note that the motivation behind using frame flipping is to be able to utilize pixels showing the motions of non-mitotic cells negatives by transforming mitosis into other events. Even if the order is flipped, the movements of the non-mitotic cell are still a non-mitotic cell feature, and we consider that these cells are effective for the training of the negative label.
Mitosis label utilization with alpha-blending pasting: Next, we paste mitosis events to the flipped pair by using copy-and-paste techniques in order to utilize the positive labels effectively. Copy and paste augmentation has been used for supervised augmentation of instance segmentation [4]. Unlike instance segmentation with object masks, we only have locations (t, x, y). A simple solution is to crop images around the mitosis position and copy and paste them to the target image, like in CutMix [23]. However, the cropped image naturally contains surrounding objects, and the generated image appears unnatural. Unnatural images cause the detection network to make biased predictions and reduce generalization performance. To avoid this problem, we propose alpha-blending pasting with a Gaussian blending mask. We blend two images by leaving the pixel value in the center and blurring the vicinity of the edge of the image.
First, we crop the image around the positive annotations and obtain a set of cropped pair and initialize and . Here, is the total number of partial annotations, while and are images before and after the mitosis of the -th annotation (Fig. 3). Define , as a cropped pair image at the -th random spatial location . We crop each image centered at to a size that is the same as that of . We update the randomly selected patch , by blending a randomly selected cropped pair with the following formula: , where is a Gaussian blending mask (Fig. 4). We generate the blending mask by blurring a binary mask around the annotation with a Gaussian filter. We use a random sigma value for the Gaussian filter. Then, we add the paste location to the set . We repeat this process random times.
2.2 Mitosis detection with generated dataset
We modified a heatmap-based cell detection method [13] to work as a mitosis detection method. Fig. 6 is an illustration of our mitosis detection model. Given two consecutive frames , the network output heatmap . We treat the channel axis as the time axis for the input. The first channel is , and the second is .
First, we generate individual heatmaps for each pasted coordinate . is defined as , where and are the coordinates of and is a hyper parameter that controls the spread of the peak. The ground truth of the heatmap at is generated by taking the maximum through the individual heatmaps, ( in Fig. 6). The network is trained with the mean square error loss between the ground truth and the output of the network . We can find the mitosis position by finding a local maximum of the heatmap.
![]() ![]() |
3 Experiments
Dataset: We evaluated our method on four datasets. The first set is HeLa [20], in which live cell images of HeLa cells expressing H2B-GFP were captured with 1100 700 resolution [20] 111We used the publicly available CTC data-set http://celltrackingchallenge.net/. We only use HeLa since the number of mitosis events in other cells is small.. Each sequence contains 92 fluorescent images with 141 mitosis events on average. The second set is ES, in which live cell images of mouse embryonic stem cells expressing H2B-mCherry were captured with 1024 1024 resolution. Each sequence contains 41 fluorescent images with 33 mitosis events on average. The third set is ES-D in which mouse embryonic stem cells expressing H2B-mCherry were induced to differentiate and used to capture live cell images. Each sequence contains 61 fluorescent images with 18 on average events on average. The fourth set is Fib, in which live cell images of mouse fibroblast cells expressing H2B-mCherry were captured with 1024 1024 resolution. Each sequence contains 42 fluorescent images with 11 mitosis events on average. Each dataset consists of four sequences of images. We performed four-fold cross-validation in which two sequences were used as training data, one as validation data, and one as test data. As shown in Fig. 6, the appearance and density are different depending on the dataset.
Implementation details: We implemented our method within the Pytorch framework [14] and used a UNet-based architecture [16] for the mitosis-detection network. The model was trained with the Adam optimizer with a learning rate of 1e-3. , which controls the spread of the heatmap, was 6. The cropping size of the positive annotations was pixels. We randomly change the number of pasting operations between and . We used random flipping, random cropping, and brightness change for the augmentation.
1-shot | 5-shot | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Method | HeLa | ES | ES-D | Fib | Ave. | HeLa | ES | ES-D | Fib | Ave. |
Baseline [13] | 0.356 | 0.581 | 0.06 | 0.235 | 0.308 | 0.591 | 0.756 | 0.277 | 0.210 | 0.459 |
GM [15] | 0.315 | 0.303 | 0.057 | 0.119 | 0.199 | 0.501 | 0.523 | 0.123 | 0.230 | 0.344 |
PU [22] | 0.030 | 0.095 | 0.012 | 0.053 | 0.048 | 0.463 | 0.538 | 0.375 | 0.224 | 0.400 |
PU-I [2] | 0.499 | 0.177 | 0.035 | 0.115 | 0.207 | 0.474 | 0.420 | 0.037 | 0.141 | 0.268 |
Ours | 0.593 | 0.740 | 0.439 | 0.440 | 0.553 | 0.795 | 0.843 | 0.628 | 0.451 | 0.610 |
Evaluation metrics: We evaluated our method using the F1 score [18], which is widely used in mitosis detection. Given ground-truth coordinates and detected coordinates, we performed one-by-one matching. If the distance of the matched pair was within spatially 15 pixels and temporally 6, we associated the closest coordinate pairs. We treated the matched pair as true positives (TP), unassociated coordinates as false positives (FP), and unassociated ground-truth coordinates as false negatives (FN).
![]() |
Comparisons: We conducted four comparisons that involved training the model with partially labeled data. For the first method, we trained the model by treating unlabeled pixels as non-mitosis ones (Baseline [13]). The second method used the Gaussian masked loss (GM [15]). The masked loss was calculated on the masked pixels around the positive-label pixels. Thus, the method ignored unlabeled pixels. The third method used positive unlabeled learning to identify mitosis from candidates obtained by the detection model trained with the masked loss (PU [22]). The fourth method generated pseudo-labels from the results of positive unlabeled learning and retrained the detection model with the pseudo-label (PU-I [2]).
In Table 1, we compared our method with previous methods in one and five-shot settings. We used N samples per sequence in the N-shot settings. For a robust comparison, we sampled one or five mitosis annotations under five seed conditions and took the average. Overall, our method outperformed all compared methods in F1 metric. GM [15], PU [22], and PU-I [2] are designed for detecting objects against simple backgrounds. Therefore, these methods are not suited to a mitosis detection task and are inferior to the baseline.
The baseline [13] treats unlabeled pixels as non-mitosis cell pixels. In the partially labeled setting, unlabeled pixels contain unannotated mitosis events, and unannotated mitosis affects performance. Unlike cell detection, mitosis detection requires identifying mitosis events from various non-mitotic cell motions, including motions that appear mitotic appearances. Although GM [15] can ignore unlabeled mitosis pixels with the masked loss, it is difficult to identify such non-mitosis motions. Therefore, GM estimates produce many false positives. PU [22] uses positive unlabeled learning to eliminate false positives from candidates obtained from the detection results with partial labels. However, positive unlabeled learning requires a positive prior in the candidates and a certain amount of randomly sampled positive samples. Since the candidates contain many false positives, the positive prior is difficult to estimate. In addition, there is no guarantee that positive unlabeled learning can work correctly with the selected N-shot annotations. Moreover, since positive unlabeled learning does not work in the mitosis detection task, PU-I [2] can not select accurate pseudo labels. Unlike these methods, our method can estimate mitosis events accurately. Since our method generates a fully labeled dataset from a partial label, it effectively uses a few partial annotations.
![]() ![]() |
Effectiveness of each module: We performed an ablation study on the HeLa dataset to investigate the effectiveness of the proposed module. We used random augmentation (i.e., random elastic transformation [17], brightness change, and gaussian noise) instead of using frame-order flipping (FOF). We generated by augmenting and input the pair to the network. In the w/o ABP setting, we directly pasted cropped images on the target image as in CutMix [23]. Table 7 demonstrates that the proposed modules improve mitosis detection performance. Fig. 9 shows examples of the estimation results for each condition. Without the FOF setting, the detection model estimates a high value for all moving cells, leading to over-detection. Without the ABP setting, the detection model overfits the directly pasted image. The directly pasted image tends to include unnatural boundaries on the edge, leading to missed detections in real images.
Robustness against missing annotations: We confirmed the robustness of the proposed method against missing annotations on the ES dataset. We changed the missing annotation rate from to . A comparison with the supervised method in terms of F1-score is shown in Fig. 7. The performance of the supervised method deteriorates as the percentage of missing labels increases, whereas the performance of the proposed method remains steady. Since our method flips the frame order, we can avoid the effects of missing annotations.
Appearance of generated dataset: Fig. 9 shows an example of the generated image pair. The cropped mitosis image pairs were pasted on the red-dotted circle. It can be seen that the borders of the original image and the pasted image have been synthesized very naturally.
4 Conclusion
We proposed a mitosis detection method using partially labeled sequences with frame-order flipping and alpha-blending pasting. Our frame-order flipping transforms unlabeled data into non-mitosis labeled data through a simple flipping operation. Moreover, we generate various positive labels with a few positive labels by using alpha-blending pasting. Unlike directly using copy-and-paste, our method generates a natural image. Experiments demonstrated that our method outperforms other methods that use partially annotated sequences on four fluorescent microscopy images.
Acknowledgements: This work was supported by JSPS KAKENHI Grant Number JP21J21810 and JST ACT-X Grant Number JPMJAX21AK, Japan.
References
- [1] Debeir, O., Van Ham, P., Kiss, R., Decaestecker, C.: Tracking of migrating cells under phase-contrast video microscopy with combined mean-shift processes. IEEE transactions on medical imaging 24(6), 697–711 (2005)
- [2] Fujii, K., Suehiro, D., Nishimura, K., Bise, R.: Cell detection from imperfect annotation by pseudo label selection using p-classification. In: MICCAI. pp. 425–434 (2021)
- [3] Gallardo, G.M., Yang, F., Ianzini, F., Mackey, M., Sonka, M.: Mitotic cell recognition with hidden markov models. In: Medical Imaging 2004: Visualization, Image-Guided Procedures, and Display. vol. 5367, pp. 661–668. SPIE (2004)
- [4] Ghiasi, G., Cui, Y., Srinivas, A., Qian, R., Lin, T.Y., Cubuk, E.D., Le, Q.V., Zoph, B.: Simple copy-paste is a strong data augmentation method for instance segmentation. In: CVPR. pp. 2918–2928 (2021)
- [5] Gilad, T., Reyes, J., Chen, J.Y., Lahav, G., Riklin Raviv, T.: Fully unsupervised symmetry-based mitosis detection in time-lapse cell microscopy. Bioinformatics 35(15), 2644–2653 (2019)
- [6] Huh, S., Bise, R., Chen, M., Kanade, T., et al.: Automated mitosis detection of stem cell populations in phase-contrast microscopy images. IEEE transactions on medical imaging 30(3), 586–596 (2010)
- [7] Liu, A.A., Tang, J., Nie, W., Su, Y.: Multi-grained random fields for mitosis identification in time-lapse phase contrast microscopy image sequences. IEEE transactions on medical imaging 36(8), 1699–1710 (2017)
- [8] Lu, Y., Liu, A.A., Chen, M., Nie, W.Z., Su, Y.T.: Sequential saliency guided deep neural network for joint mitosis identification and localization in time-lapse phase contrast microscopy images. IEEE Journal of Biomedical and Health Informatics 24(5), 1367–1378 (2019)
- [9] Mao, Y., Yin, Z.: A hierarchical convolutional neural network for mitosis detection in phase-contrast microscopy images. In: MICCAI. pp. 685–692 (2016)
- [10] Mao, Y., Yin, Z.: Two-stream bidirectional long short-term memory for mitosis event detection and stage localization in phase-contrast microscopy images. In: MICCAI. pp. 56–64 (2017)
- [11] Nie, W.Z., Li, W.H., Liu, A.A., Hao, T., Su, Y.T.: 3d convolutional networks-based mitotic event detection in time-lapse phase contrast microscopy image sequences of stem cell populations. In: CVPR. pp. 55–62
- [12] Nishimura, K., Bise, R.: Spatial-temporal mitosis detection in phase-contrast microscopy via likelihood map estimation by 3dcnn. In: EMBC. pp. 1811–1815 (2020)
- [13] Nishimura, K., Wang, C., Watanabe, K., Bise, R., et al.: Weakly supervised cell instance segmentation under various conditions. Medical Image Analysis 73, 102182 (2021)
- [14] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: An imperative style, high-performance deep learning library. NeurIPS 32 (2019)
- [15] Qu, H., Wu, P., Huang, Q., Yi, J., Yan, Z., Li, K., Riedlinger, G.M., De, S., Zhang, S., Metaxas, D.N.: Weakly supervised deep nuclei segmentation using partial points annotation in histopathology images. IEEE transactions on medical imaging 39(11), 3655–3666 (2020)
- [16] Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: MICCAI. pp. 234–241 (2015)
- [17] Simard, P.Y., Steinkraus, D., Platt, J.C., et al.: Best practices for convolutional neural networks applied to visual document analysis. In: ICDAR. vol. 3 (2003)
- [18] Su, Y.T., Lu, Y., Liu, J., Chen, M., Liu, A.A.: Spatio-temporal mitosis detection in time-lapse phase-contrast microscopy image sequences: A benchmark. IEEE Transactions on Medical Imaging 40(5), 1319–1328 (2021)
- [19] Thirusittampalam, K., Hossain, M.J., Ghita, O., Whelan, P.F.: A novel framework for cellular tracking and mitosis detection in dense phase contrast microscopy images. IEEE journal of biomedical and health informatics 17(3), 642–653 (2013)
- [20] Ulman, V., Maška, M., Magnusson, K.E., Ronneberger, O., Haubold, C., Harder, N., Matula, P., Matula, P., Svoboda, D., Radojevic, M., et al.: An objective comparison of cell-tracking algorithms. Nature methods 14(12), 1141–1152 (2017)
- [21] Yang, F., Mackey, M.A., Ianzini, F., Gallardo, G., Sonka, M.: Cell segmentation, tracking, and mitosis detection using temporal context. In: MICCAI. pp. 302–309 (2005)
- [22] Yang, Y., Liang, K.J., Carin, L.: Object detection as a positive-unlabeled problem. BMVC (2020)
- [23] Yun, S., Han, D., Oh, S.J., Chun, S., Choe, J., Yoo, Y.: Cutmix: Regularization strategy to train strong classifiers with localizable features. In: ICCV. pp. 6023–6032 (2019)