Issue |
A&A
Volume 691, November 2024
|
|
---|---|---|
Article Number | A331 | |
Number of page(s) | 23 | |
Section | Numerical methods and codes | |
DOI | https://doi.org/10.1051/0004-6361/202349113 | |
Published online | 25 November 2024 |
CLAP
I. Resolving miscalibration for deep learning-based galaxy photometric redshift estimation
1
Pengcheng Laboratory,
Nanshan District,
Shenzhen,
Guangdong
518000,
P.R. China
2
Aix-Marseille Université, CNRS/IN2P3, CPPM,
Marseille
13009,
France
3
School of Physics and Microelectronics, Zhengzhou University,
Zhengzhou,
Henan
450001,
P.R. China
4
Department of Physics E. Pancini, University Federico II,
Via Cinthia 6,
80126
Naples,
Italy
5
School of Physics and Astronomy, Sun Yat-sen University,
Zhuhai Campus, 2 Daxue Road, Tangjia,
Zhuhai,
Guangdong
519082,
P.R.China
6
CSST Science Center for Guangdong-Hong Kong-Macau Great Bay Area,
Zhuhai,
Guangdong
519082,
P.R. China
7
Department of Astronomy, The Ohio State University,
Columbus,
OH
43210,
USA
8
Center for Cosmology and AstroParticle Physics (CCAPP), The Ohio State University,
Columbus,
OH
43210,
USA
9
Research School of Astronomy & Astrophysics, Australian National University,
Cotter Rd.,
Weston,
ACT 2611,
Australia
10
School of Computing, Australian National University,
Acton,
ACT 2601,
Australia
★ Corresponding author; linqf@pcl.ac.cn
Received:
28
December
2023
Accepted:
22
September
2024
Obtaining well-calibrated photometric redshift probability densities for galaxies without a spectroscopic measurement remains a challenge. Deep learning discriminative models, typically fed with multi-band galaxy images, can produce outputs that mimic probability densities and achieve state-of-the-art accuracy. However, several previous studies have found that such models may be affected by miscalibration, an issue that would result in discrepancies between the model outputs and the actual distributions of true redshifts. Our work develops a novel method called the Contrastive Learning and Adaptive KNN for Photometric Redshift (CLAP) that resolves this issue. It leverages supervised contrastive learning (SCL) and k-nearest neighbours (KNN) to construct and calibrate raw probability density estimates, and implements a refitting procedure to resume end-to-end discriminative models ready to produce final estimates for large-scale imaging data, bypassing the intensive computation required for KNN. The harmonic mean is adopted to combine an ensemble of estimates from multiple realisations for improving accuracy. Our experiments demonstrate that CLAP takes advantage of both deep learning and KNN, outperforming benchmark methods on the calibration of probability density estimates and retaining high accuracy and computational efficiency. With reference to CLAP, a deeper investigation on miscalibration for conventional deep learning is presented. We point out that miscalibration is particularly sensitive to the method-induced excessive correlations among data instances in addition to the unaccounted-for epistemic uncertainties. Reducing the uncertainties may not guarantee the removal of miscalibration due to the presence of such excessive correlations, yet this is a problem for conventional methods rather than CLAP. These discussions underscore the robustness of CLAP for obtaining photometric redshift probability densities required by astrophysical and cosmological applications. This is the first paper in our series on CLAP.
Key words: methods: data analysis / techniques: image processing / surveys / galaxies: distances and redshifts
© The Authors 2024
Open Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This article is published in open access under the Subscribe to Open model. Subscribe to A&A to support open access publication.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.