Morphological classification of astronomical images with limited labelling
Abstract
The task of morphological classification is complex for simple parameterization, but important for research in the galaxy evolution field. Future galaxy surveys (e.g. EUCLID) will collect data about more than a galaxies. To obtain morphological information one needs to involve people to mark up galaxy images, which requires either a considerable amount of money or a huge number of volunteers. We propose an effective semi-supervised approach for galaxy morphology classification task, based on active learning of adversarial autoencoder (AAE) model. For a binary classification problem (top level question of Galaxy Zoo 2 decision tree) we achieved accuracy 93.1% on the test part with only 0.86 millions markup actions, this model can easily scale up on any number of images. Our best model with additional markup achieves accuracy of 95.5%. To the best of our knowledge it is a first time AAE semi-supervised learning model used in astronomy.
1Moscow State University CMC, Moscow, Russia; [email protected]
2Space Research Institute of RAS, Moscow, Russia; [email protected]
1 Introduction
Galaxies exhibit a wide variety of shapes, colors and sizes. These properties are indicative of their age, formation conditions, and interactions with other galaxies over the course of many Gyr. Studies of galaxy formation and evolution use morphology to probe the physical processes that give rise to them. Traditionally, morphological classification has been carried out by professional astronomers using visual inspections of images obtained from telescopes. The number of detected galaxies in modern sky surveys and computer simulations has been found to be growing rapidly. The number of images in surveys (SDSS, DESI LIS, PanSTARRS) is billions of galaxies, future surveys such as Euclid (Refregier et al. 2010) will provide images of more than billion of galaxies. Manually inspecting all these images to annotate them with morphological information is impractical for either individual astronomers or small teams.
An attempt to solve this problem was the Galaxy Zoo project by Willett et al. (2013): volunteers classified images of galaxies through a web interface. Participants were asked various questions such as "How rounded is the galaxy?" or "Is there a convex in the center?" Depending on the answer, it was determined which question would be asked next. The questions formed a decision tree that covered a wide range of morphological categories. The crowdsourcing project also faces the problem of a large amount of data: if we try to study the morphology of only 1% of the volume of images from Euclid, we will get millions of objects for morphological classification. A project of this magnitude would take over sixty years to process all the data at the pace of the Galaxy Zoo 2.
There are approaches to galaxy classification based on the ideas of active learning. They allow to increase the classification speed by 8 times at the expense of a slight loss of quality, but even at this marking rate, it will take 7.5 years for the SWAP by Beck et al. (2018) (SoTA model) to handle 1% of all Euclid data.
The main objectives of our research: 1) create a model for the morphological classification of galaxies, 2) the model should use small as possible amount of labeled data, 3) predictive ability should be better than current SoTA solutions, 4) use of active learning technique. The second task is the main one, since reducing the number of required markup images will reduce the number of primary classification actions carried out by volunteers / research groups. This, in turn, will significantly speed up the marking of morphological annotations of modern surveys.
2 Data
We chose Kaggle Galaxy Zoo dataset for our experiments. With 138000 images of galaxies in jpg format. As mentioned above, Galaxy Zoo is a crowdsourcing project, each image in it was classified based on the votes of volunteers, an average of 42 people per galaxy. Thus, we have a dependence between amount of labeled data and markup actions: 1 labeled image markup actions. Annotation of data was provided by Kyle Willett, KGZ organizer. Authors are grateful for his help and kind words.
3 Methods
In this work, we propose semi-supervised algorithm, modernization of the adversarial autoencoder by Makhzani et al. (2016). We did classification over a continuous latent representation using additional MLP layer. We trained encoder in semi-supervised mode with both labeled and unlabeled data, the second was an order of magnitude more. In the process of training adversarial autoencoder, in addition to the classical "uniform" selection of training samples, we also used active learning approach. The training examples were selected sequentially, based on the prediction of the model trained in the previous step with fewer examples. At the first step, 4% of the total number of images were randomly marked as a training sample, model was trained on them, other images were classified with that model, and the next 1% of the total number of images was marked, for which algorithm predicted the probability closest to 0.5 (random prediction). The model was retrained, process continued until the total amount of labeled data was equal to 10% of all training data. We approximated the results to the dataset size used by Beck et al. (2018). To do this, we took the accuracy obtained on the unlabeled part and transferred it to an enlarged one in accordance with the size of the SWAP dataset minus 20k labeled images and calculated the final results as a composition of these values.

4 Results
The main learning outcomes: 1) Semi-supervised model (AAE) achieves an accuracy of 91% on 10% labeled data (850,000 markup actions for 20k images) of all training data, which is quite close to the result of a similar SWAP model (93.1% on 936000 markup actions); 2) Active learning improves prediction, AAE+AL model trained on the same amount of labeled images achieves an accuracy of 93.1%. This result can be easily scaled up on any amount of images without additional markup cost; 3) If we estimate the accuracy over the entire Galaxy Zoo 2 dataset (226124 images as in Beck et al. (2018)), including labeled examples (as the Beck et al. (2018) of SWAP did ), our model achieves an accuracy of 95.5%, which is quite close to the best SWAP result (95.7%). At the same time, it requires 2.5 times less markup actions (850000 vs. 2300000 for SWAP).

Model | Acc.(%) | Galaxies | Markup actions | Add markup? |
---|---|---|---|---|
SWAP | 95.7 | 226124 | 2298772 | Yes |
AAE+AL | 95.5 | 226124 | 854310 | Yes |
SWAP+RF | 93.1 | 210803 | 936887 | Yes |
AAE+AL (20k sample) | 93.1 | Any | 854310 | No |
Let be the size of the entire dataset to be classified, A be the limit of labeled images used in active learning. Let’s denote by - the ratio of to : . We measured how the prediction accuracy of the model changes with a fixed and growing . We sampled data with the range from 30k to 130k with equal to 7k and 20k. We note that the best accuracy of classification is achieved with more labeled data. So if we increase the total amount of data, but keep the value R, we get the best results (Figures 3).

5 Conclusion
To the best of our knowledge its a first time AAE used in astronomy. Our model achieves accuracy 93.1% on 20k labeled data without additional markup cost, this model can easily scale up on any number of images. Our best model with additional markup achieves accuracy of 95.5%, while using 2.5 times less markup actions than Beck et al. (2018).
References
- Beck et al. (2018) Beck, M. R., Scarlata, C., Fortson, L. F., Lintott, C. J., Simmons, B. D., Galloway, M. A., Willett, K. W., Dickinson, H., Masters, K. L., Marshall, P. J., & et al. 2018, Monthly Notices of the Royal Astronomical Society, 476, 5516–5534. URL http://dx.doi.org/10.1093/mnras/sty503
- Makhzani et al. (2016) Makhzani, A., Shlens, J., Jaitly, N., Goodfellow, I., & Frey, B. 2016, Adversarial autoencoders. 1511.05644
- Refregier et al. (2010) Refregier, A., Amara, A., Kitching, T. D., Rassat, A., Scaramella, R., & Weller, J. 2010, Euclid imaging consortium science book. 1001.0061
- Willett et al. (2013) Willett, K. W., Lintott, C. J., Bamford, S. P., Masters, K. L., Simmons, B. D., Casteels, K. R. V., Edmondson, E. M., Fortson, L. F., Kaviraj, S., Keel, W. C., & et al. 2013, Monthly Notices of the Royal Astronomical Society, 435, 2835–2860. URL http://dx.doi.org/10.1093/mnras/stt1458