Issue |
A&A
Volume 664, August 2022
|
|
---|---|---|
Article Number | A81 | |
Number of page(s) | 18 | |
Section | Numerical methods and codes | |
DOI | https://doi.org/10.1051/0004-6361/202142952 | |
Published online | 10 August 2022 |
O’TRAIN: A robust and flexible ‘real or bogus’ classifier for the study of the optical transient sky★
1
Université Paris-Saclay, CNRS, CEA, Astrophysique, Instrumentation et Modélisation de Paris-Saclay,
91191
Gif-sur-Yvette, France
e-mail: damien.turpin@cea.fr
2
Ecole Centrale Lille,
Cité scientifique CS 20048,
59651
Villeneuve d’Ascq cedex, France
3
Sorbonne Université, CNRS, UMR 7095, Institut d’Astrophysique de Paris,
98 bis boulevard Arago,
75014
Paris, France
4
CEICO, Institute of Physics of the Czech Academy of Sciences,
Prague, Czech Republic
5
Instituto de Astrofísica de Andalucía (IAA-CSIC),
Glorieta de la Astronomía s/n,
18008
Granada, Spain
6
IRAP, Université de Toulouse, CNRS, UPS,
14 Avenue Edouard Belin,
31400
Toulouse, France
7
Université Paul Sabatier Toulouse III, Université de Toulouse,
118 route de Narbonne,
31400
Toulouse, France
Received:
19
December
2021
Accepted:
20
April
2022
Context. Scientific interest in studying high-energy transient phenomena in the Universe has risen sharply over the last decade. At present, multiple ground-based survey projects have emerged to continuously monitor the optical (and multi-messenger) transient sky at higher image cadences and covering ever larger portions of the sky every night. These novel approaches are leading to a substantial increase in global alert rates, which need to be handled with care, especially with regard to keeping the level of false alarms as low as possible. Therefore, the standard transient detection pipelines previously designed for narrow field-of-view instruments must now integrate more sophisticated tools to deal with the growing number and diversity of alerts and false alarms.
Aims. Deep machine learning algorithms have now proven their efficiency in recognising patterns in images. These methods are now used in astrophysics to perform different classification tasks such as identifying bogus from real transient point-like sources. We explore this method to provide a robust and flexible algorithm that could be included in any kind of transient detection pipeline.
Methods. We built a convolutional neural network (CNN) algorithm in order to perform a ‘real or bogus’ classification task on transient candidate cutouts (subtraction residuals) provided by different kinds of optical telescopes. The training involved human-supervised labelling of the cutouts, which are split into two balanced data sets with ‘true’ and ‘false’ point-like source candidates. We tested our CNN model on the candidates produced by two different transient detection pipelines. In addition, we made use of several diagnostic tools to evaluate the classification performance of our CNN models.
Results. We show that our CNN algorithm can be successfully trained on a large and diverse array of images on very different pixel scales. In this training process, we did not detect any strong over- or underfitting with the requirement of providing cutouts with a limited size no larger than 50 × 50 pixels. Tested on optical images from four different telescopes and utilising two different transient detection pipelines, our CNN model provides a robust ‘real or bogus’ classification performance accuracy from 93% up to 98% for well-classified candidates.
Key words: methods: numerical / techniques: image processing
The codes and diagnostic tools presented in this paper are available at https://github.com/dcorre/otrain
© K. Makhlouf et al. 2022
Open Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This article is published in open access under the Subscribe-to-Open model. Subscribe to A&A to support open access publication.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.