Issue |
A&A
Volume 664, August 2022
|
|
---|---|---|
Article Number | A4 | |
Number of page(s) | 17 | |
Section | Numerical methods and codes | |
DOI | https://doi.org/10.1051/0004-6361/202142463 | |
Published online | 03 August 2022 |
Finding strong gravitational lenses through self-attention
Study based on the Bologna Lens Challenge
1
National Centre for Nuclear Research,
Warsaw, Poland
e-mail: Hareesh.Thuruthipilly@ncbj.gov.pl; Adam.Zadrozny@ncbj.gov.pl; Agnieszka.Pollo@ncbj.gov.pl
2
Jagiellonian University,
Kraków, Poland
3
Department of Astronomy, Beijing Normal University,
100875, PR
Beijing, China
Received:
16
October
2021
Accepted:
11
May
2022
Context. The upcoming large-scale surveys, such as the Rubin Observatory Legacy Survey of Space and Time, are expected to find approximately 105 strong gravitational lenses by analysing data many orders of magnitude larger than those in contemporary astronomical surveys. In this case, non-automated techniques will be highly challenging and time-consuming, if they are possible at all.
Aims. We propose a new automated architecture based on the principle of self-attention to find strong gravitational lenses. The advantages of self-attention-based encoder models over convolution neural networks (CNNs) are investigated, and ways to optimise the outcome of encoder models are analysed.
Methods. We constructed and trained 21 self-attention-based encoder models and five CNNs to identify gravitational lenses from the Bologna Lens Challenge. Each model was trained separately using 18000 simulated images, cross-validated using 2000 images, and then applied to a test set with 100 000 images. We used four different metrics for evaluation: classification accuracy, the area under the receiver operating characteristic (AUROC) curve, and TPR0 and TPR10 scores (two metrics of evaluation for the Bologna challenge). The performance of self-attention-based encoder models and CNNs participating in the challenge are compared.
Results. The encoder models performed better than the CNNs. They were able to surpass the CNN models that participated in the Bologna Lens Challenge by a high margin for the TPR0 and TPR10. In terms of the AUROC, the encoder models with 3 × 106 parameters had equivalent scores to the top CNN model, which had around 23 × 106 parameters.
Conclusions. Self-attention-based models have clear advantages compared to simpler CNNs. They perform competitively in comparison to the currently used residual neural networks. Self-attention-based models can identify lensing candidates with a high confidence level and will be able to filter out potential candidates from real data. Moreover, introducing the encoder layers can also tackle the overfitting problem present in the CNNs by acting as effective filters.
Key words: gravitational lensing: strong / methods: data analysis / techniques: image processing / cosmology: observations
© H. Thuruthipilly et al. 2022
Open Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This article is published in open access under the Subscribe-to-Open model. Subscribe to A&A to support open access publication.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.