Open Access
Issue
A&A
Volume 674, June 2023
Article Number A161
Number of page(s) 15
Section Astronomical instrumentation
DOI https://doi.org/10.1051/0004-6361/202245778
Published online 20 June 2023

© The Authors 2023

Licence Creative CommonsOpen Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

This article is published in open access under the Subscribe to Open model. Subscribe to A&A to support open access publication.

1 Introduction

Radio astronomy is at the cusp of a revolution in terms of the sensitivity that can be achieved at centimeter and meter wavelengths. The various radio-astronomical interferometers such as the LOw-Frequency ARray (LOFAR; van Haarlem et al. 2013), MeerKAT radio telescope (Jonas & MeerKAT Team 2016), Australian square kilometer array pathfinder (ASKAP; Hotan et al. 2021), and Murchison Widefield Array (MWA; Tingay et al. 2013) are already producing promising results. At the same time, radio astronomy is an extremely data-intensive science. It is expected that the upcoming projects will produce data volumes on an exabyte scale (Scaife 2020). Thus, it will be very challenging for astronomers to undertake standard radio data analysis tasks such as calibration, imaging and source localization using traditional techniques. Hence, it is vital to design fast and efficient techniques to replace the traditional data analysis approaches in radio astronomy.

At millimeter (mm) wavelengths the Atacama Large Millimeter Array (ALMA) has led to several large imaging and spectroscopic programs thanks to its excellent sensitivity. Of particular interest in the context of the current work are the various large programs targeting extra-galactic deep fields. This includes, for example, the Reionization Era Bright Emission Line Survey (REBELS; Bouwens et al. 2022), the ALMA SPECtroscopic Survey in the Hubble Ultra-Deep Field (ASPECS; Walter et al. 2016), the ALMA Large Program to INvestigate [CII] at Early times (ALPINE; Le Fèvre et al. 2020; Béthermin et al. 2020; Faisst et al. 2020), and the GOODS-ALMA survey at 1.1 mm (Franco et al. 2018). Another rich ALMA data set on which our current study is focused is the Automated Mining of the ALMA Archive in the COSMOS Field (A3COSMOS) data set1 (Liu et al. 2019). These surveys study the gas and dust properties of galaxies at high redshifts.

A key technical component while analyzing these data is to accurately identify sources (their positions, fluxes, and sizes) in noise-limited images produced by radio interferometers. This is important, for example, when calculating the number density and luminosity function of astronomical sources. It has important implications for constraining various physical models in astrophysics and cosmology. Traditionally, radio source properties were measured using a two-dimensional (2D) Gaussian fit to the light profile (Condon 1997) and usually required some level of manual intervention. Such manual interventions become increasingly difficult when dealing with large-area ALMA surveys mentioned above. Also with the all-sky radio surveys such as the Faint Images of the Radio Sky at Twenty-cm (FIRST) survey (Becker et al. 1995) and NRAO Very Large Array (VLA) Sky Survey (NVSS; Condon et al. 1998), which contain millions of sources, manual source detection is difficult. Future large-area radio surveys (Norris et al. 2013) are expected to detect increase in the number of sources by an order of magnitude (e.g., the Evolutionary Map of the Universe (EMU) survey; Norris et al. 2011). Radio surveys of extra-galactic deep fields using upgraded radio facilities, such as the VLA-COSMOS 3 GHz survey (Smolčić et al. 2017) and the MeerKAT International GHz Tiered Extra-galactic Exploration (MIGHTEE) survey (Jarvis et al. 2016; Heywood et al. 2022), are already detecting several thousands of sources in a single field.

Thus, it is essential to build an automatic source detection algorithm that would accurately detect sources above the noise level with high completeness and simultaneously have a low number of false detections. There are several automatic source detection tools available in the literature, for example, Search and Destroy (SAD)2, Source-Extractor (SExtractor; Bertin & Arnouts 1996), AEGEAN (Hancock et al. 2012, 2018) and Python Blob Detection and Source Finder (PyBDSF; Mohan & Rafferty 2015), PySE (Carbone et al. 2018), CAESAR (Riggi et al. 2016, 2019), PROFOUND (Robotham et al. 2018; Hale et al. 2019), and SOFIA (Serra et al. 2015; Westmeier et al. 2021). Hopkins et al. (2015) provide a detailed comparison between different source detection algorithms available in the literature and their limitations. This works very well for bright sources. However, it is much more interesting to detect faint sources, typically with a signal-to-noise ratio (S/N) below 5.0, with high level of completeness and purity. This not only offers the ability to probe sources at even higher redshifts than currently achieved. It also has the potential to detect new kinds of astronomical sources lying close to the noise level, which may have been missed by studies using traditional source detection techniques.

Recent advances in machine learning and in particular deep neural networks in the form of convolutional neural networks (CNNs) have led to a lot of success in radio astronomy. In particular, CNNs have been extensively used to classify radio galaxy morphologies (e.g., Aniyan & Thorat 2017; Lukic et al. 2019a; Ma et al. 2019; Tang et al. 2019; Bowles et al. 2021; Riggi et al. 2023). In particular, Riggi et al. (2023) have used Mask R-CNN object detection framework on the ASKAP EMU survey data to perform both object detections and classifications. Schmidt et al. (2022) used CNNs designed for super-resolution applications directly on the UV data to up-sample features in the case of sparse sampling, for instance, very-long-baseline interferometry. The images produced from these UV data show good recovery of source properties. Also, CNNs have shown great promise in point source detection, as demonstrated in ConvoSource (Lukic et al. 2019b) and DeepSource (Vafaei Sadr et al. 2019). DeepSource is shown to be perfect in terms of purity and completeness down to a S/N of 4 and outperforms the current state-of-the-art source detection algorithm PyBDSF in several metrics. Recently, it was shown that an encoder-decoder based neural network DECORAS (Rezaei et al. 2022) can perform source detection even on dirty images down to a S/N of 5.5 and can also recover various source properties such as fluxes and sizes quite accurately. Delli Veneri et al. (2023) further shows that deep learning-based source detection can also be performed on dirty spectral data cubes and shows a good recovery of source properties such as morphology, flux density, and projection angle.

A common problem in most of the source-finding algorithms in the literature is that they are performed on the image plane and mostly on CLEAN images, which are computationally expensive to produce and scale poorly with the data volumes. Thus, even if these source finding algorithms are made efficient, they will still be limited by the time taken to produce the CLEAN images. It is well known that CLEAN leads to several imaging artifacts that can affect the purity of these source-finding algorithms. And despite automating several steps in source finding, these approaches still require some amount of manual intervention to exclude imaging artifacts. In this work, we circumvent these problems by designing a novel direct uv-plane based source localization algorithm using recent advances in deep neural networks.

Motivated by the achievements of the deep neural networks (DNN) made in many domains, we propose a fast and efficient DNN-based framework for source localization. In general, this framework takes as input a low-dimension vector of sampled uv-observations and outputs the source position on sky in the form of a binary map. Our proposed framework is targeted towards source detections in ALMA continuum images at mm wavelengths, particularly with the aim of detecting low S/N sources and to speed up the source detection process overall. We trained and tested our proposed framework on simulated ALMA data. We investigated the impact of different factors on the performance of the proposed framework, namely, the representation of complex-valued input data via real and imaginary or magnitude and phase real-valued components, the impact of receiver noise and atmospheric noise due to the presence of water vapor, the impact of the S/N values of the sources, and the impact of the number of sources in the field of view. We provide a detailed analysis of the execution time of the proposed framework and compare it with those of the traditional source localization pipeline.

The paper is organized as follows. The ALMA data simulation procedure, the chosen parameters and the analysis are given in Sect. 2. The traditional pipeline of source localization is described in Sect. 3. The framework proposed in this work is explained in Sect. 4. Section 5 is dedicated to the analysis of the obtained results and Sect. 6 offers comparisons with the literature. Finally, we present our conclusions in Sect. 7.

2 Data

We train the proposed DNN-based framework on synthetic data with known source localization. All reported experiments have been performed exclusively on synthetic data. In the future, we plan to apply the proposed frameworks to the real observations taken from the (A3COSMOS) data set (Liu et al. 2019). Thus, our simulated ALMA observations are designed to somewhat match the A3COSMOS observations.

For all our simulations, we used the Common Astronomy Software Applications (CASA) data processing software V6.2 (McMullin et al. 2007; CASA Team 2022). We produced simulations for the 12-m ALMA array with a fixed configuration from the ALMA cycle 5.3. All our simulation pointings were randomly distributed within a radius of 1 deg around the COSMOS field centered on: J2000 10h00m28.6s + 02d12m21.0s.

For simplicity, we fixed the ALMA observing band to Band-6 centered at 230 GHz split in 240 channels with a channel width of 7.8 MHz. Our simulation pipeline follows the standard approach of first simulating a true sky model at a known phasecentre with known source positions and fluxes. The sky model consists only of sources with a Gaussian light profile with sizes close to or below the synthesized beam. In each pointing, the source position, size (i.e., the major and minor axis of the Gaussian), position angle, S/N, and total number of sources (in the range between 1 and 5) are kept at random and drawn from an uniform distribution. This range in the number of sources per pointing and their sizes are chosen from true data from the A3COSMOS data in Band-6 for the ALMA compact configuration (similar to ours). Instead of drawing our simulation parameters exactly from the observed distribution in A3COSMOS data, we used a uniform distribution but the range of these parameters is motivated by the A3COSMOS data set (see the discussion in the following paragraph). The sky model contains a 512 × 512 pixel image with a pixel size of 0.1″. This sky model is used to generate the noise free uv-data using the SIMALMA task (McMullin et al. 2007). We sample each visibility every 10 s for a total of 20 min of integration on-source and choose the hour angle range such that the observing field is at transit. Appendix A shows a typical UV sampling for one of the simulations and the corresponding dirty beam. We then added the ALMA receiver and atmospheric noise due to water vapor to the uv-data using the PWV parameter value of 1796 (McMullin et al. 2007). This is a typical value in Band-6 as mentioned in the ALMA technical handbook3. We then average the visibilities, both noisy and noise-free, in time and frequency, by gridding the visibilities on a uniform grid using the MSUVBIN task (McMullin et al. 2007). Finally, we produce the dirty and CLEAN images from both the noisy and noise-free visibilities using the standard TCLEAN task (McMullin et al. 2007). For our simulation setup, we typically reach a rms noise of 50 µJy in our dirty and CLEAN images. The size of the primary and synthesized beam is 22.86″ and 0.82″, respectively. Table summarizes the various simulation parameters.

Although geared towards real observations from A3COSMOS, our simulations are intentionally chosen to not match them exactly. In particular, our sources are assigned random on-sky positions, fluxes, and sizes – as opposed to choosing directly from the distribution of sources from true observations or those motivated by cosmological simulations. The phase center and number of sources in each simulation pointing are also randomly distributed around the COSMOS field. This ensures that during training we do not end up in the region of latent space of our deep-learning model, which is only trained to work for a particular data set. Thus, it does not learn additional patterns which might exist in real observations or cosmological simulations such as, source clustering, the luminosity function, size distribution etc. Instead, our model is trained to be flexible enough to be applied to any other new ALMA data. On the other hand, for our current work, we fix the ALMA observing band, channel width and total integration time. This is to reduce the number of free parameters in ALMA simulations. In the future, we will test the effect of changing these two parameters especially when we apply our model to real data.

We produced a total of 9164 independent simulations or pointings with the sky model containing between 1 and 5 sources. In total, we ended up with 27632 sources across all the pointings. Each simulation produces ~3.5GB of data which includes the uv-data and dirty and CLEAN images that is ~35 TB for the entire data set4. The simulations are produced on the LESTA-computing cluster. The more details are given in the Appendix B.

Figures 1 and 2 show examples of the simulated data. The red circles highlight the positions of the sources in the true sky models. It is important to mention that the generated data set includes the same challenges encountered in real data, for example, the presence of blended sources (as in Fig. 2) and sources that are hardly distinguishable from the background noise as in Fig. 1 (top-middle and bottom-right). For the training of our framework, we perform the train-validation-test splitting three times with different seeds. Table 2 shows the number of sky models with different numbers of sources in the test subsets.

The notation we use is defined as follows: x ∊ ℝN×M denotes the true sky model, where N and M are the image size, set to 512 in our simulations. Then, x^dirtyN×M${{{\bf{\hat x}}}_{dirty}} \in {^{N \times M}}$ and x^cleanN×M${{{\bf{\hat x}}}_{{\rm{clean}}}} \in {^{N \times M}}$ denote dirty and CLEAN image, respectively, y ∊ ℂK stands for the sampled uv-data with K = 14005. m ∊ {0,1}N×M is a binary source map, in which source pixels are set to 1, and background pixels are set to 0. We calculated the S/N as: S/N=total fluxσnoise,${S \mathord{\left/ {\vphantom {S N}} \right. \kern-\nulldelimiterspace} N} = {{{\rm{total}}\,{\rm{flux}}} \over {{\sigma _{{\rm{noise}}}}}},$(1)

where the total flux is the true source intensity and σnoise is the noise rms. As explained in Table 1, σnoise ~ 50 µJy.

Table 1

Summary of the data simulation parameters using CASA.

thumbnail Fig. 1

Examples of simulated data used in this study. Top-left: true sky model. Top-middle: noise-free model. Top-right: noisy. Bottom-left: true sky model. Bottom-middle: noise-free. Bottom right: noisy. For both noise-free and noisy cases, the CLEAN representations are visualized. The sources are highlighted as red circles.

thumbnail Fig. 2

Example of closely located sources of different intensity. For better visibility, the cropped and zoomed parts of the true sky models are shown.

Table 2

Number of sky models with different numbers of sources in the test subsets.

3 Traditional pipeline of source localization

The traditional pipeline of source localization is based on the reconstruction of the dirty and CLEAN images. The schematic representation of the traditional pipeline is given in Fig. 3.

3.1 Dirty image recovering

The recovering of the dirty image might be formulated as a recovering of a high dimension sky model6 x ∊ ℝN·M from the corresponding low-dimensional sampled uv-visibility y ∊ ℂK (KN·M) corrupted by noise e: y=Wx+e,${\bf{y}} = {\bf{Wx}} + {\bf{e}},$(2)

where W = PΩΨ ∊ ℂK×N·M is a measurement sub-sampling matrices with an orthonormal basis of Ψ ∊ ℂN·M×N·M and sampling operator of PΩ : ℂN·M K and |Ω| = K.

It should be pointed out that due to the physical imaging constraints in the radio-astronomy, Ψ corresponds to the Fourier operator and Pω is determined by the antennas configuration, measured frequencies, Earth's movement, sampling, and integration time.

The recovering of the x^dirtyN×M${{{\bf{\hat x}}}_{{\rm{dirty}}}} \in {^{N \times M}}$ consists in: (i) PΩTy${\bf{P}}_{\rm{\Omega }}^T{\bf{y}}$ -expanding the observation y to a N × M representation by placing zeros in the entries corresponding to ΩC that is the complementary support set of Ω, (ii) adding the corresponding symmetrical signal via S (.) related to the symmetry property of Fourier and then (iii) applying the inverse Fourier Ψ*: x^dirty=Ψ*(PΩTy+S(PΩTy)).${{\bf{\hat x}}_{{\rm{dirty}}}} = {{\bf{\Psi }}^*}\left( {{\bf{P}}_{\rm{\Omega }}^T{\bf{y}} + S\left( {{\bf{P}}_{\rm{\Omega }}^T{\bf{y}}} \right)} \right).$(3)

thumbnail Fig. 3

Schematic representation of the traditional pipeline.

thumbnail Fig. 4

Schematic representation of the proposed framework.

3.2 Clean image recovering

Due to the presence of missing Fourier frequencies in PΩC${{\bf{P}}_{{{\rm{\Omega }}^C}}}$, the dirty image contains a lot of artifacts, leading to an increase in the false detected sources. In order to remove these artifacts, the CLEAN algorithm (Högbom 1974) is applied to the dirty image.

The CLEAN fundamental method is used in radio astronomy and consists of several steps. First, it finds the intensity and position of the peak that is of the greatest absolute intensity in the dirty image x^dirty${{\bf{\hat x}}_{{\rm{dirty}}}}$ Second, it generates at this position a spike of an intensity equal to the product of a damping factor and the intensity at that position. Usually, the damping factor is ≤ 1 and is termed the loop gain. The generated spikes are convolved with the instrumental point source function (PSF). Then, the obtained instrumental response is subtracted from the dirty image x^dirty${{\bf{\hat x}}_{{\rm{dirty}}}}$· This procedure is repeated unless any remaining peak is below some user-specified level. The search for peaks may be constrained to specified areas of the image, called CLEAN windows. Then the accumulated point sources are convolved with an idealized CLEAN beam that, usually, is an elliptical Gaussian fitted to the central lobe of the dirty beam. Finally, the remaining residual of the dirty image is added. The obtained reconstruction is called CLEAN image x^clean${{\bf{\hat x}}_{{\rm{clean}}}}$·

In terms of the fast processing of large amounts of data, CLEAN as well as its accelerated versions such as the w-Stacking Clean (WSCLEAN; Offringa et al. 2014) act as the main bottleneck. The reconstruction remains time-consuming since several parameters have to be adjusted in iterative runs.

3.3 Source localization

Traditionally, the source localization is applied to the CLEAN images. Depending on the antenna configuration and the corresponding amount of artifacts, it might also be applied to the dirty images. Nowadays, a broad range of different source localization approaches exist that are used in the traditional pipeline. Hopkins et al. (2015) provide a good overview of these methods. In our work, we focus on PyBDSF (Mohan & Rafferty 2015).

4 Proposed framework

We propose a DNN-based framework that performs the source localization in the form of a binary map m ∊ {0,1}N×M directly from the uv-data by taking only sampled visibility data without reconstruction of dirty or CLEAN images as an input. The general scheme of the basic framework is illustrated in Fig. 4 and consists of three steps: (1) input data pre-processing (i.e., normalization), (2) DNN-processing: stage 1 and 2, (3) post-binarization and source localization.

It should be pointed out that the sampled visibility data y ∊ ℂK are complex values. However, modern DNNs are not designed to work with complex values. To deal with this, we decompose the complex-valued uv-samples y ∊ ℂK into real-valued real and imaginary or magnitude and phase representation ∊ ℝK . Due to the nature of the Fourier transform, in the case of real and imaginary representation, the trained model gθ at the stage 1 is additive in nature, as shown in on the left side of Fig. 5. While in case of the magnitude and phase, it is multiplicative in nature as shown on the right side of Fig. 5. The other steps are universal and remain unchanged7.

thumbnail Fig. 5

Detailed architecture of the gθ model, where ȳ denotes the normalized input data . Left panel: ȳ ∊ ℝK is represented by the real and imaginary components. Right panel: ӯ ∊ ℝK is represented by the magnitude and phase components.

4.1 Input data pre-processing

Examples of the raw noise-free and noisy input data, ∊ ℝK, are given in Fig. 6. As can be seen from the noise-free case, the real and imaginary components have the same dynamic range of power 10−4 (Fig. 6, top-left), while the magnitude and phase have different dynamic range (Fig. 6, top-right): the magnitude is of a power of 10−4, similarly to the real and imaginary components, and the phase is in a much wider range, going from −3 to 3. Regarding the noisy case, it is important to note that in the real, imaginary, and magnitude components, the noise dominates and increases the dynamic range of data by two orders of magnitude, while the phase component is less affected by noise and preserves the same dynamic range: from −3 to 3. To guarantee stable DNN training and to avoid the vanishing of gradients, the real and imaginary components are normalized by multiplying by 1000 and clipped in the range [−10, 10]. The clipping allows us to reduce the impact of strong outliers. The magnitude component was multiplied by 100 and clipped in the range of [0, 1]8. The phase component was processed without any normalization. Empirically, this type of normalization was found to be optimal during the proposed model training. The normalized vector is denoted as ӯ ∊ ℝK

4.2 DNN processing

It is important to highlight that our goal is to perform the source localization. Thus, we are not interested, for the time being, in the prediction of the sources fluxes and other parameters such as size and so on. In this respect, stages 1 and 2 are trained to minimize the similarity score with respect to the true binary source map, m ∊ {0, 1}N×M.

4.2.1 Stage 1 : Real and imaginary representation

For the real and imaginary representation of the input data the schematic architecture of the model, gθ is shown in the left panel of Fig. 59. At first, the given input low-dimensional representation ӯ ∊ ℝK is mapped into a higher dimension representation aı ∊ ℝN·M via a fully connected layer. Then the obtained high dimension representation is reshaped into a square representation a2 ∊ ℝN×M Finally, the weighted sum of the obtained components produces the output m^st.1N×M${{\bf{\hat m}}_{st.1}} \in {^{N \times M}}$.

4.2.2 Stage 1 : Magnitude and phase representation

For the magnitude and phase representation of the input data, the schematic architecture of the model gθ is shown in right Fig. 510. First, we performed the weighted element-wise multiplication of the two low-dimensional components in ӯ ∊ ℝK. Then, the resulting low-dimensional representation, b1 ∊ ℝK, was mapped into a higher dimensional representation, b2 ∊ ℝN·M, via a fully connected layer. Finally, this high-dimensional representation is reshaped to a square representation that corresponds to the output m^st.1N×M${{\bf{\hat m}}_{st.1}} \in {^{N \times M}}$ of the stage 1, which is a real valued estimation of the source map.

The estimation of the parameters θ of the trained model gθ is based on solving the minimization problem: θ^=argminθst.1(θ)=mse(m,m^st.1),${\bf{\hat \theta }} = \mathop {\arg \min }\limits_{\bf{\theta }} {L_{st.1}}\left( {\bf{\theta }} \right) = {L_{{\rm{mse}}}}\left( {{\bf{m}},{{{\bf{\hat m}}}_{st.1}}} \right),$(4)

where m^st.1=gθ(y¯)${{\bf{\hat m}}_{st.1}} = {g_{\bf{\theta }}}\left( {{\bf{\bar y}}} \right)$ and ℒmse(.,.) denotes the mean square loss.

4.2.3 Stage 2

Due to the fact that the network prediction, m^st.1N×M${{\bf{\hat m}}_{st.1}} \in {^{N \times M}}$, is valued as real, the model gθ preserves some information about the true source intensity. Therefore, some predicted sources might be of low intensity, as shown in Fig. 7 (the second sub-figure). As a result, after the binarization process, certain sources might be lost, especially in the noisy case. In this respect, stage 2 might be considered as a quality enhancement stage for the source detection where (as can be seen in Fig. 7, i.e., the third sub-figure) the predicted source intensity is close to the binary representation. Taking into account the simple nature of the expected prediction in the form of a simple binary map without any complex shapes and textures, we used a simple auto-encoder model11 as the model qϕ. For the more complicated tasks such as source intensity or any other estimation of physical parameters, the model qϕ might be represented by more advanced models, such as UNet (Long et al. 2015) or Transformers (Vaswani et al. 2017).

The estimation of the parameters ϕ of the trained model qϕ is done by solving the minimization problem: ϕ^=argminϕst.2(ϕ)=mse(m,m^st.2),$\hat \phi = \mathop {\arg \min }\limits_\phi {L_{st.2}}\left( \phi \right) = {L_{{\rm{mse}}}}\left( {{\bf{m}},{{{\bf{\hat m}}}_{st.2}}} \right),$(5)

where m^st.2=qϕ(gθ*(y¯))${{\bf{\hat m}}_{st.2}} = {q_\phi }\left( {{g_{{{\bf{\theta }}^*}}}\left( {{\bf{\bar y}}} \right)} \right)$, θ* denotes fixed pre-trained model parameters and ℒmse(.,.) denotes the mean square loss.

thumbnail Fig. 6

Examples of input data. Top-left: real and imaginary components in the noise-free case. Top-right: magnitude and phase components in the noise-free case. Bottom-left: real and imaginary components in the noisy case. Bottom-right: magnitude and phase components in the noisy case.

thumbnail Fig. 7

Example of the proposed model’s outputs. Left to right: example of the true binary source map m ∊ {0,1}N×M. Proposed framework estimations, m^st.1N×M${{{\bf{\hat m}}}_{st.1}} \in {^{N \times M}}$ and m^st.2N×M${{{\bf{\hat m}}}_{st.2}} \in {^{N \times M}}$ based on the real and imaginary input data representation (for the magnitude and phase representation the results look similar). Final estimation: m^{ 0,1 }N×M${\bf{\hat m}} \in {\left\{ {0,1} \right\}^{N \times M}}$

4.3 Post-binarization and source localization

Taking into account a need to satisfy the differentiability of DNN for the gradient propagation at the training, the DNN cannot produce the binary outputs. In this respect, the post-binarization stage is necessary. However, instead of performing the hard thresholding that might produce source blobs of an uncontrollable size, a morphological-based binarization was used.

The morphological binarization consists of determining the optimal threshold value12 for every particular DNN output, m^N×M${\bf{\hat m}} \in {^{N \times M}}$, on the fly. After the binarization, the connected neighborhoods are detected, while overly small or big regions are rejected. Finally, the source position was estimated for each detected region by taking their centroids.

4.4 Conceptual advantages

Compared to the traditional pipeline, conceptually, the proposed framework has the following advantages. The input data dimensionality is much smaller: only the sampled uv-data are processed to produce a localization binary map, m, while the traditional pipeline works with the full-size uv-plane, where missing frequencies are filled with zeros. Moreover, the proposed framework might be extended to any type of data sources and even for mixed types, while the CLEAN method used in the traditional pipeline offers only a poor reconstruction of regions of extended emission. For this reason, it is usually applied to the sky models that are only composed of point sources. For regions of extended emission, other methods are used. In this respect, the traditional pipeline is more demanding in terms of the expert knowledge. Finally, the CLEAN reconstruction as well as the dirty image estimation are not required as such to carry out the source detection, which represents a considerable computational advantage in practice.

5 Results and analysis

By default and unless specified, all results are given as an average over three test subsets mentioned in Sect. 2. For simplicity, we named the proposed framework based on the real and imaginary components of the input data representation as “Re & Im” and those based on the magnitude and phase components are given as “Mag & Phase.”

5.1 Metrics

We cross-matched detected sources with true sources in each sky model using a match radius, R, equal to the beam size as shown in Fig. 8 to compute the model performance metrics listed below. The impact of the matched radius in the results is discussed in Sect. 5.3.

To evaluate the performance of the traditional pipeline and the proposed framework, we used the purity and completeness metrics that are defined as follows.

Purity shows the fraction of the true sources among all detected sources: Purity=TPTP+FP,${\rm{Purity}} = {{{\rm{TP}}} \over {{\rm{TP}} + {\rm{FP}}}},$(6)

where TP and FP denote true positive and false positive sources, respectively. Here, 1 – purity represents the fraction of false detections.

Completeness is equal to the fraction of true sources that are successfully detected: Completeness=TPTP+FN${\rm{Completeness}} = {{{\rm{TP}}} \over {{\rm{TP}} + {\rm{FN}}}}$(7)

where FN represents the false negative (missing) sources.

thumbnail Fig. 8

Schematic representation of the acceptable distance of R between the true and predicted sources.

thumbnail Fig. 9

Purity and completeness of the methods under investigation in the noise-free case. Small black lines on the top of each bar correspond to the standard deviation from the average.

5.2 Results on noise-free data

We consider the noise-free case as an ideal condition measurement. In real observations, this reflects a hypothetical situation, since all measurements are corrupted by noise of different nature. However, we found it to be important to validate the performance of the investigated approaches under the assumption that there is no noise to understand their baseline performance.

Figure 9 demonstrates the purity and completeness obtained for PyBDSF applied to the CLEAN and dirty images and the proposed framework in four different configurations. We chose the source detection parameters in such a way as to have the maximum completeness under the assumption that the acceptable purity should be about 94%13. It is easy to see that PyBDSF on the CLEAN and dirty images and the proposed Re & Im framework exhibit a similar performance. There is no big difference in the performance of PyBDSF for the CLEAN and dirty images. The Re & Im stage 1 and 2 provide very close results. In the case of Mag & Phase, the performance is worse. As it is mentioned in Sect. 4.1, there is a big difference in the dynamic range of the magnitude and phase components. On one side, such a difference is natural and should be preserved. On the other side, it is a dis-advantage and a challenge for the DNN training. This explains the obtained non-optimal results.

thumbnail Fig. 10

Difference between the noise-free (left) and noisy (right) prediction of the proposed Re & Im stage 1 for a simulated sky model.

thumbnail Fig. 11

Purity and completeness of the methods under investigation in the noisy case, averaged over all S/N values. The small black lines on the top of each bar correspond to the standard deviation from the average.

5.3 Results on noisy data

The noisy scenario is of a particular interest for our study of real observations. In this respect, we provide a more detailed analysis of the obtained results.

Before presenting the results of the proposed framework for the noisy case, we would like to underline the difference in the predictions between the noise-free and noisy case. It helps gain an understanding of why the proposed framework is better than the traditional pipeline. The noise-free and noisy predictions of the Re & Im after stage 1 are shown in Fig. 10. It is important to highlight that in contrast to the CLEAN or dirty noisy data used in the traditional pipeline (Fig. 1), the proposed framework predictions based on the noisy input data are free from the background noise, which is very important for the efficient source localization.

In Fig. 11, we show the purity and completeness for PyBDSF and the proposed framework in four different configurations. For a fair comparison, the parameters of source localization for all methods under investigation are selected in such a way to have a purity of about 90%14. We fixed and used these parameters for all the following experiments.

In comparison to the noise-free case, in the noisy case we have slightly smaller purity (about 3–4% less), but the obtained completeness is significantly smaller for all methods under investigation. Secondly, it is important to note that the PyBDSF completeness does not exceed 25 – 26%, which is very small for practice. At the same time, it is interesting to note that the proposed Mag & Phase framework that has the worst performance in the noise-free case does, in fact, outperform the state-of-the-art PyBDSF on the noisy data. Its completeness is about 60% after stage 1 and about 55% after stage 2. The proposed Re & Im framework demonstrates the best completeness about 69% after stage 1 and 74% after stage 2. However, it should be pointed out that stage 1 is more stable, as shown by the lower standard deviation.

It is obvious that the performance of the source localization highly depends on the S/N value. In Fig. 12, we plot the completeness obtained for each method under investigation with respect to the S/N of the source. We can see that when S/N ≤ 3 PyBDSF cannot detect many sources (the completeness does not exceed 2–3%). Then, for S/N~ 2, the proposed Mag & Phase reaches 10% and for S/N~ 3 it already has a 25% completeness. The performance of the proposed Re & Im is even better: for S/N~ 2 the completeness is about 20% and for S/N ~ 3 it is about 50% after stage 1 and about 55% after stage 2. With the increasing S/N, the performance of all methods increases. However, for the maximum S/N the state-of-the-art PyBDSF can get the completeness only up to about 75%, while the proposed Re & Im achieves 93 – 94%. It is also important to note that the Re & Im stage 2 offers improvement for S/Ns of 2–5, as compared to the Re & Im stage 1. For the very low or high S/N, the performance at both stages is similar.

As mentioned in Sect. 2, one of the challenges of our data set is closely located sources that are difficult to distinguish, especially in the noisy case with increasing source densities. It is obvious that when we have more sources in the sky model, the probability to have close sources is higher. We analyze the performance of PyBDSF and the proposed Re & Im with respect to the source S/N specifically for different numbers of sources in the sky models. The derived results are shown in Fig. 13. In general, the obtained behavior is similar to that shown in Fig. 12. However, there are several interesting facts that we can take note of. First of all, the behavior of PyBDSF is stable and does not depend too much on the number of sources in the sky model. On the other hand, when the number of sources increases the efficiency of the proposed framework decreases, especially for the low S/N values. For example, for S/N ~ 2 in the case with only one source (Fig. 13, left panel), the completeness is about 50%, while in the case with five sources (Fig. 13 the right panel), the completeness drops down to about 15 – 20%. With the S/N increase, the drop in efficiency slows down. This phenomenon is naturally connected with the increased entropy of multi-source spectra. Keeping the fixed number of samples in the uv-plane, the sky model with several sources is characterized by a lower entropy spectrum. Therefore, there is a fundamental accuracy limit of source detection under the restricted sparse sampling of the uv-plane. This is one of the directions for our future investigation. At the same time, it is important to point out that for the setup under investigation and the chosen antenna configuration (for mode details, see Sect. 2), the expected number of sources ranges from one to five. For the larger number of sources, the ambiguity in the under-sampled uv-data increases. For this reason, the configuration with the larger number of antennas should be used. Under the chosen setup, even though the completeness of Re & Im drops, its general efficiency is much better than PyBDSF.

Another important question that we study is the dependency of source localization accuracy on the distance to the true sources. In Fig. 8, we schematically show the locations of the true source and the corresponding predicted source. We consider the predicted source as true if it is inside a given radius, R, around the true source. In the current experiments, we set an R = α · beam size15 where α is a scale factor. The results shown in Fig. 14 are obtained for the test subset 116. It is important to mention that in all previous experiments α = 1 or, in other words, the radius R equals to the beam size. In Fig. 14, we can observe some increase in the purity when α > 1, but the saturation is achieved already for α = 1.2–1.5. For the completeness, the saturation is achieved for α = 1. In Fig. 15, we plot the mean distance in arcsec between the centers of true and detected sources. The gray bars show the relative number of detected sources. The relative means were divided by the total number of true sources for the given S/N. It should be pointed out that in the case of the Re & Im framework, the deviation from the mean value is quite stable and does not depend a great deal on the number of detected sources or S/N. At the same time, with the increase in the S/N and the number of detected sources, the mean value convergences to about 0.16″ after stage 1 and to about 0.19″ after stage 2. For PyBDSF, we can see quite small deviation from the mean values for S/N = 2 and a large deviation for S/N = 3. This can be explained by the very small number of detected sources and, as a consequence, the poor statistics. With the increase in the S/N and the number of detected sources, the deviation from the mean value decreases and the mean value convergences to about 0.15″. It is important to note that Re & Im framework reaches the convergence at S/N = 5, while for PyBDSF we can observe the convergence only after S/N = 8.

thumbnail Fig. 12

Dependence of completeness on S/N of the sources.

5.4 Execution time

To investigate the question of the time complexity, we measured the execution time of the source localization for the methods under investigation. The obtained CPU time in seconds17 is summarized in Table 3. For the proposed framework, the column labeled “reconstruction from uv” corresponds to the DNN-processing, Sect. 4.2. The “source localization” column corresponds to the procedure explained in Sect. 4.3. It should be pointed out that the recovery of the dirty image takes about 14 s. The recovery of the CLEAN image takes about 17 s. On the other hand, the proposed framework in the slowest case (Mag & Phase after two stages) estimates the real-valued source map only in 0.33 s. The source localization by PyBDSF, compared to the most efficient Re & Im stage 2 is about 4.3 times slower. In terms of the total time complexity, PyBDSF on the CLEAN data (the best traditional pipeline results) is ~38 times slower than the proposed Re & Im stage 2 (the best results) – even on the CPU execution.

thumbnail Fig. 13

Dependence on the efficiency of PyBDSF and the proposed Re & Im framework on the number of sources in the sky model. Left: one source. Middle: three sources. Right: five sources.

Table 3

Execution time for the source localisation per sky model (noisy data).

thumbnail Fig. 14

Dependence on α, where α=Rbeam size$\alpha = {R \over {{\rm{beam}}\,{\rm{size}}}}$ and R is the distance between the true and predicted sources, in the noisy case, averaged over all S/N values.

thumbnail Fig. 15

Mean distance (the y/-axis) between the true and predicted sources in the noisy case. The coloured semi-transparent background shows the standard deviation from the mean value. The gray bars show the relative number of detected sources.

6 Discussion

6.1 Performance comparisons with the literature

In classical approaches to interferometric observations, fidelity and completeness are two important measures of the significance and “statistical importance” resulting from source detections and characterizations, which are generally determined by empirical methods. For example, the fidelity is computed by comparing the number of sources detected with positive flux to those in the negative image (Fidelity=1NnegNpos)$\left( {{\rm{Fidelity}} = 1 - {{{N_{neg}}} \over {{N_{pos}}}}} \right)$ to empirically determine a S/N threshold above which individual sources are considered “reliably detected” (see, e.g., Aravena et al. 2016; Béthermin et al. 2020). Furthermore, completeness is generally determined using Monte Carlo source injections in the observational data (e.g., in the image plane).

Since fidelity requires a flux measurement (which is not included in the proposed framework), it is not possible to compare the performance of our method for this quantity, as already mentioned above (Sect. 5.1). However, we have compared our results to those of Béthermin et al. (2020), who presented a detailed analysis of the completeness from extra-galactic ALMA continuum observations with properties similar to those of our sky simulations. Following Béthermin et al. (2020), we therefore determined the completeness as a function of the normalized injected flux: fnorm=total fluxσnoisebminbmajbmin2+smin2bmaj2+smaj2,${f_{{\rm{norm}}}} = {{{\rm{total}}\,{\rm{flux}}} \over {{\sigma _{{\rm{noise}}}}}} \cdot {{{b_{\min }} \cdot {b_{{\rm{maj}}}}} \over {\sqrt {b_{\min }^2 + s_{\min }^2} \cdot \sqrt {b_{{\rm{maj}}}^2 + s_{{\rm{maj}}}^2} }},$(8)

where bmin and bmaj correspond to the minor and major axes of the beam, and smin and smaj denote the minor and major axes of the source. This quantity resembles an effective, normalized S/N, and it encapsulates in particular variations of sources sizes and provides a simple functional description of the completeness, as shown by Béthermin et al. (2020).

The results obtained on our simulated data are shown in Fig. 16. First, we note that the behavior of PyBDSF on our CLEAN and dirty images is similar to what is seen for the find_peak source detector from astropy (Astropy Collaboration 2013) used by Béthermin et al. (2020), especially at fnorm ≲ 4. Most importantly, for all normalized fluxes fnoгm ≲ 5 the proposed framework achieves significantly higher completeness than classical methods. Finally, for high normalized fluxes (and high S/N), the completeness obtained by the proposed framework and Béthermin et al. (2020) is similar. From this comparison, we conclude that the proposed method is not only expected to significantly speed up, but also to strongly improve source detection in interferometric imaging.

thumbnail Fig. 16

Dependence of completeness on normalized S/N compared to Béthermin et al. (2020).

6.2 Impact of the noise

As mentioned in Sect. 2, in the ALMA setup under investigation, there are two sources of the noise. The first one is the ALMA receiver noise that is fixed in our case, since we fixed the total integration time and observing band. The second one is the atmospheric noise related to the water vapor. To investigate the sensitivity of the proposed framework to the change of noise statistics, we consider the change of the atmospheric noise by varying the PWV parameter. We chose ten different PWV values from 0.472 to 5.186. These values were chosen from the ALMA technical handbook18 and represent the typical values observed at the ALMA site. For each PWV value, we simulated 100 test samples. Then we tested the proposed Re & Im framework without retraining and without any change in the test parameters. We performed the comparison with PyBDSF that is also used without any change of parameters for a fair comparison. The obtained purity and completeness are shown in Fig. 17. It is important to notice that, on average, for the Re & Im framework, we can observe the same dynamic as represented in Sect. 5.3; namely, stage 2 is slightly superior to stage 1. The performance of PyBDSF under the noise PWV smaller than in the main simulations (i.e., PWV = 1.796) is similar to the one reported in Sect. 5.3. With the increase in the noise we can observe a drastic decrease in purity from 98–100% to 30% and to 3% for the PWV = 5.186. This shows the need for expert knowledge in adapting the parameters. For the proposed Re & Im framework, there is also decrease in the performance but it is not so drastic. On average, the purity remains about 87–90%, while the completeness decreases from about 60–65% to 45–47% – this is 10% better than the results from PyBDSF.

In addition, we investigated an extreme case that involves the performance obtained on the noise-free data (PWV = 0) with the models trained on the noisy data. The obtained results are shown in Fig. 17 (bottom right). In general, the behavior of the proposed Re & Im framework is quite good. One can observe a certain decrease in the performance at stage 2, while the purity and completeness obtained at stage 1 are high. The completeness obtained by PyBDSF is smaller than in case of noisy data. This can be explained by the choice of parameters, assuming the presence of noise.

6.3 Caveats

It should be pointed out that, despite the advantages discussed above, the proposed framework also has certain limitations. Although it requires less expert knowledge, the amount of training data increases with increasing data complexity, such as the complexity of the shape of the source and their variability, the increase in the number of sources in the field of view, the proximity of the sources with different S/N.

Moreover, taking into account that the proposed framework shows high sensitivity to the low S/N sources, it might lead to false sources in the source free sky models (e.g., pure background noise). And although such a scenario usually is not considered among state-of-the-art approaches, we tested the proposed Re & Im framework on 1000 source free noisy sky models. After stage 1, the false sources are detected in about 40% of the sky models, while after stage 2, only in 27% of cases. The results obtained can be explained by the fact that the source free sky models were not taken into account during the framework training.

6.4 Future developments and applications

The proposed framework can be applied to real data, and will be tested with available data from the ALMA archive, such as data from A3COSMOS (Liu et al. 2019). This will, in particular, also allow us to examine the behavior with data taken in different conditions and accounting for all noise sources present in the system.

The next tasks to be tackled from uv-data alone will include source characterization that are measurements of fluxes, source morphologies, and others, as well as the treatment of extended sources, possibly with complex morphologies. New machine learning-based approaches on visibilities should also be able to handle spectral lines and thus be applicable to an even broader range of astrophysical questions. In parallel, our framework should also be tested and generalized to cover a wide range of interferometric data taken with very different facilities and across a wide spectral range from the millimeter to the radio domain. If proven successful, machine learning based methods could have a strong impact on our future handling of interferometric data and help push the discovery space of upcoming observatories even further.

thumbnail Fig. 17

Purity and completeness of PyBDSF on CLEAN data and the proposed Re & Im framework for the different noise levels, averaged over all S/Ns. From top left to bottom right: PWV = 0.472; PWV = 0.55; PWV = 0.658; PWV = 0.72; PWV = 0.913; PWV = 1.05; PWV = 1.262; PWV = 2.748; PWV = 3.9; PWV = 5.186; PWV = 0.

7 Conclusions

Radio astronomy is at a historic moment in its development. Innovative DNN-based methods may drastically change the way measurements are processed and, thus, may also improve our capacity to detect faint and complex signals. In this context, we have taken a new ML-based approach to solve the problem of source localization, directly working on the natural measurement sets (visibilities) and without reconstructing the dirty or CLEAN image. The proposed framework consists of two stages generating source localization maps (see Fig. 4).

To train the network and then validate and test the proposed framework, we used synthetic data generated with the CASA data processing software (McMullin et al. 2007), which is the software tool for interferometric data/observations with ALMA and similar observatories, and PyBDSF (Mohan & Rafferty 2015) for the classical source detection. The sky simulations generated in this work were chosen to represent typical extra-galactic (sub)-millimeter continuum survey observations undertaken with ALMA. The simulations included both ideal noise-free and realistic noisy simulations. The comparisons between the proposed DNN-based approach and existing state-of-the-art source localization pipelines can be summarized as follows:

  • In ideal noise-free cases, the proposed framework has overall the same purity and completeness as the traditional pipeline using PyBDSF source localization;

  • For noisy data, the proposed DNN-based method fares significantly better than classical source localization algorithms; For example, the Re & Im framework shows completeness that is more than three times as high as it is for the same purity as the traditional PyBDSF-based pipeline for all sky models, when considering a uniform, random distribution of sources with S/N = 1–10 (Fig. 12);

  • In the low-S/N regime (S/N < 5), the performance gain of the proposed method is very high: while the traditional pipeline achieves a completeness of 2–5% for S/N < 3, the proposed framework detects sources with a completeness of ~20 (45–55)% for S/N =2 (3);

  • The new source-detection method represents an important gain in execution time, with total execution times that are more than 30 times faster than the traditional pipeline that involves image reconstruction from the uv-plane and source detection (Table 3).

We have investigated the impact of different factors on the efficiency of the proposed framework, in particular:

  • Input data representation: While traditional DNNs are designed to work with real-valued data, we have tested representations of the complex-valued uv-data by real-valued real and imaginary or magnitude and phase components. The obtained results show that the real and imaginary components are better better suited to the proposed framework, since they have the same dynamic range in contrast to the magnitude-phase case;

  • Source density: Traditional source detection in the image plane demonstrates a stable behavior that does not depend much on the number of sources. However, in the proposed framework the completeness decreases with increasing numbers of sources (Fig. 13). Despite this, in tested conditions which represent typical ALMA extra-galactic continuum surveys, the proposed framework reaches a higher completeness than traditional source detection algorithms.

In short, we have developed a DNN-based method of source detection using only uv-plane observations. We have shown that it provides strong improvements in detecting sources in the low-to-intermediate-S/N regime. The new approach can already be applied to existing interferometric observations and it opens many new possibilities which will be explored in the near future. Machine learning-based methods have the potential to significantly alter our approach to interferometric data, as we enter the era of new facilities like the Square Kilometer Array19 (SKA).

Acknowledgements

We acknowledge the referee for their comments. O. Taran and O. Bait are supported by the AstroSignals Sinergia Project funded by the Swiss National Science Foundation. We also thank B. Magnelli and M. Béthermin, and the nordic arc node https://nordic-alma.se.

Appendix A UV Coverage and the dirty beam

Figure A.1 shows the typical UV coverage for our simulations and the corresponding dirty beam. Note: the dirty beam has several sidelobes, but they are of a smaller amplitude, owing to the relatively filled UV.

thumbnail Fig. A.1

Example of the total UV coverage for one of our ALMA 12-m array simulation pointing lasting 20 mins with 240 channels (top). The corresponding dirty beam is shown in the bottom panel.

Appendix B LESTA computing cluster and simulation details

The simulations presented in this work are produced on the LESTA computing cluster hosted at the University of Geneva Department of Astronomy. The details of the used resources:

  • 16 Intel Xeon (E5-2640 v3, 2.6GHz) with 128GB (2133 MHz) memory (per node)

    • number of nodes: 32;

    • number of cores: 512

  • 32 Intel(R) Xeon(R) Gold 5218 CPU @ 2.30GHz with 256GB (2666 MHz) memory (per node):

    • number of nodes: 8;

    • number of cores: 256

Each simulation or pointing was run in parallel using slurm job arrays on both resources, depending on availability. Each simulation takes approximately 12 minutes.

Appendix C Sub-sampling

It is important to note that networks are quite demanding for input data stability such as the data size, the data dynamic range, and so on. In the considered antenna configuration, there are 50 antennas. That gives 1225 measurements. Taking into account the Earth’s movement during the integration time, along with the source positions and the subsequent gridding, the final number of measured frequencies (i.e., dimensionality of the input data) increases and is variable from about 1300 until about 1500. To solve the issue with the variable size of the input data, we chose 1400 most frequently measured uv-positions. To do this, we first took 100 gridded data at random and converted them to the binary maps, where unmeasured frequencies were set to 0 and the measured ones to 1, disregarding the real intensity. Then we summed up these 100 binary maps and chose 1400 positions with the highest values sorted in the decreasing order. That, in its turn, leads to the final binary map with 1400 values equal to 1 and the others equal to 0. Then every gridded piece of data was multiplied by this final binary map. It is clear that in the cases when the gridded data contain more than 1400 measured frequencies, it leads to the lost of information. However, we allowed for this in order to have the fixed size input data for the network.

Appendix D Proposed framework training

The Python implementation of the proposed framework is be publicly available at https://github.com/taranO/ml-based_source_localization_from_uv-plane.

For the training of the proposed framework, the data set was split into the train (80%), validation (10%), and test subsets (10%). The training was done three times under three different seeds.

Stage 1

Table D.1

Architecture of the gθ model for ȳ ∊ ℝK represented by the real and imaginary components.

Table D.2

Architecture of the gθ model for ȳ ∊ ℝK represented by the magnitude and phase components.

Training of the proposed framework was performed on the NVIDIA Titan RTX GPU with 24GB memory during 1000 epochs (the training time is about 3 hours) with the learning rate 1e-4 and the batch size of 312.

Stage 2 training was performed on the NVIDIA GeForce RTX 2080 Ti GPU with 11GB memory during 500 epochs (the training time is about 8 hours) with the learning rate 1e-3 and the batch size of 32.

Table D.3

Architecture of the qϕ model.

Post-binarization: we do not apply the fixed threshold t to binarize the DNN outputs m^N×M${{\bf{\hat m}}^{ \in N \times M}}$. Instead, the binarization threshold t is chosen on the fly for each particular m^${\bf{\hat m}}$. From the prior knowledge, we know that there are no more than five sources in each simulation. To determine the optimal threshold, t, the proposed method performs an iterative thresholding until the number of detected sources, p, both true and false, is not bigger than 15. This value was chosen empirically based on the trade-off between the corresponding purity and completeness as shown in Fig. D.1.

thumbnail Fig. D.1

Dependence of the purity and completeness on the limit on the number of detected sources p.

Appendix E State-of-the-art source detectors

E.1 Noise-free data

Table E.1

Performance (in%) of the state-of-the-art source detectors on the CLEAN images in the noise-free case.

A comparison of the performance of the state-of-the-art PyBDSF, AEGEAN (Hancock et al. 2012, 2018), and PySE (Carbone et al. 2018) methods on our data set is given in Table E.1. The parameters of the detectors were chosen to provide the purity of about 94%. It is easy to see that the purity and completeness obtained for PyBDSF and AEGEAN are very similar. Under the same purity, PySE has slightly smaller completeness.

E.2 Noisy data

Table E.2

Performance (in%) of the state-of-the-art source detectors on the CLEAN images in the noisy case, averaged over all S/Ns.

Similarly to the noise-free case, we compare the performance of the state-of-the-art detectors PyBDSF, AEGEAN, and PySE. The obtained purity and completeness are given in Table E.2. The parameters of the detectors were chosen to provide the maximum purity under the assumption that the completeness should be no less than 25%. In contrast to the noise-free case, in case of noisy data under the same completeness PySE outperforms AEGEAN in terms of the purity. AEGEAN detector has almost two times smaller purity compared to PyBDSF. PyBDSF demonstrates the best performance.

References

  1. Aniyan, A. K., & Thorat, K. 2017, ApJS, 230, 20 [Google Scholar]
  2. Aravena, M., Decarli, R., Walter, F., et al. 2016, ApJ, 833, 71 [NASA ADS] [CrossRef] [Google Scholar]
  3. Astropy Collaboration (Robitaille, T. P., et al.) 2013, A&A, 558, A33 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  4. Becker, R. H., White, R. L., & Helfand, D. J. 1995, ApJ, 450, 559 [Google Scholar]
  5. Bertin, E., & Arnouts, S. 1996, A&AS, 117, 393 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  6. Béthermin, M., Fudamoto, Y., Ginolfi, M., et al. 2020, A&A, 643, A2 [Google Scholar]
  7. Bouwens, R. J., Smit, R., Schouws, S., et al. 2022, ApJ, 931, 160 [NASA ADS] [CrossRef] [Google Scholar]
  8. Bowles, M., Scaife, A. M. M., Porter, F., Tang, H., & Bastien, D. J. 2021, MNRAS, 501, 4579 [CrossRef] [Google Scholar]
  9. Carbone, D., Garsden, H., Spreeuw, H., et al. 2018, Astron. Comput., 23, 92 [NASA ADS] [CrossRef] [Google Scholar]
  10. CASA Team (Bean, B., et al.) 2022, PASP, 134, 114501 [NASA ADS] [CrossRef] [Google Scholar]
  11. Condon, J. J. 1997, PASP, 109, 166 [NASA ADS] [CrossRef] [Google Scholar]
  12. Condon, J. J., Cotton, W. D., Greisen, E. W., et al. 1998, AJ, 115, 1693 [Google Scholar]
  13. Delli Veneri, M., Tychoniec, E., Guglielmetti, F., Longo, G., & Villard, E. 2023, MNRAS, 518, 3407 [Google Scholar]
  14. Faisst, A. L., Schaerer, D., Lemaux, B. C., et al. 2020, ApJS, 247, 61 [NASA ADS] [CrossRef] [Google Scholar]
  15. Franco, M., Elbaz, D., Béthermin, M., et al. 2018, A&A, 620, A152 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  16. Hale, C., Robotham, A., Davies, L., et al. 2019, MNRAS, 487, 3971 [NASA ADS] [CrossRef] [Google Scholar]
  17. Hancock, P. J., Murphy, T., Gaensler, B. M., Hopkins, A., & Curran, J. R. 2012, MNRAS, 422, 1812 [Google Scholar]
  18. Hancock, P. J., Trott, C. M., & Hurley-Walker, N. 2018, PASA, 35, e011 [Google Scholar]
  19. Heywood, I., Jarvis, M. J., Hale, C. L., et al. 2022, MNRAS, 509, 2150 [Google Scholar]
  20. Högbom, J. 1974, A&ASS, 15, 417 [Google Scholar]
  21. Hopkins, A. M., Whiting, M. T., Seymour, N., et al. 2015, PASA, 32, e037 [Google Scholar]
  22. Hotan, A. W., Bunton, J. D., Chippendale, A. P., et al. 2021, PASA, 38, e009 [NASA ADS] [CrossRef] [Google Scholar]
  23. Jarvis, M., Taylor, R., Agudo, I., et al. 2016, in MeerKAT Science: On the Pathway to the SKA, 6 [Google Scholar]
  24. Jonas, J., & MeerKAT Team. 2016, in MeerKAT Science: On the Pathway to the SKA, 1 [Google Scholar]
  25. Le Fèvre, O., Béthermin, M., Faisst, A., et al. 2020, A&A, 643, A1 [Google Scholar]
  26. Liu, D., Lang, P., Magnelli, B., et al. 2019, ApJS, 244, 40 [Google Scholar]
  27. Long, J., Shelhamer, E., & Darrell, T. 2015, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 3431 [Google Scholar]
  28. Lukic, V., Brüggen, M., Mingo, B., et al. 2019a, MNRAS, 487, 1729 [Google Scholar]
  29. Lukic, V., de Gasperin, F., & Brüggen, M. 2019b, Galaxies, 8, 3 [NASA ADS] [CrossRef] [Google Scholar]
  30. Ma, Z., Xu, H., Zhu, J., et al. 2019, ApJS, 240, 34 [NASA ADS] [CrossRef] [Google Scholar]
  31. McMullin, J. P., Waters, B., Schiebel, D., Young, W., & Golap, K. 2007, ASP Conf. Ser., 376, 127 [Google Scholar]
  32. Mohan, N., & Rafferty, D. 2015, Astrophysics Source Code Library [record ascl:1502.007] [Google Scholar]
  33. Norris, R. P., Hopkins, A. M., Afonso, J., et al. 2011, PASA, 28, 215 [Google Scholar]
  34. Norris, R. P., Afonso, J., Bacon, D., et al. 2013, PASA, 30, e020 [NASA ADS] [CrossRef] [Google Scholar]
  35. Offringa, A., McKinley, B., Hurley-Walker, N., et al. 2014, MNRAS, 444, 606 [Google Scholar]
  36. Rezaei, S., McKean, J. P., Biehl, M., & Javadpour, A. 2022, MNRAS, 510, 5891 [NASA ADS] [CrossRef] [Google Scholar]
  37. Riggi, S., Ingallinera, A., Leto, P., et al. 2016, MNRAS, 460, 1486 [NASA ADS] [CrossRef] [Google Scholar]
  38. Riggi, S., Vitello, F., Becciani, U., et al. 2019, PASA, 36, e037 [NASA ADS] [CrossRef] [Google Scholar]
  39. Riggi, S., Magro, D., Sortino, R., et al. 2023, Astron. Comput., 42, 100682 [NASA ADS] [CrossRef] [Google Scholar]
  40. Robotham, A., Davies, L., Driver, S., et al. 2018, MNRAS, 476, 3137 [NASA ADS] [CrossRef] [Google Scholar]
  41. Scaife, A. M. M. 2020, Philos. Trans. Roy. Soc. Lond. A, 378, 20190060 [NASA ADS] [Google Scholar]
  42. Schmidt, K., Geyer, F., Fröse, S., et al. 2022, A&A, 664, A134 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  43. Serra, P., Westmeier, T., Giese, N., et al. 2015, MNRAS, 448, 1922 [Google Scholar]
  44. Smolčić, V., Novak, M., Bondi, M., et al. 2017, A&A, 602, A1 [Google Scholar]
  45. Tang, H., Scaife, A. M. M., & Leahy, J. P. 2019, MNRAS, 488, 3358 [Google Scholar]
  46. Tingay, S. J., Goeke, R., Bowman, J. D., et al. 2013, PASA, 30, e007 [Google Scholar]
  47. Vafaei Sadr, A., Vos, E. E., Bassett, B. A., et al. 2019, MNRAS, 484, 2793 [CrossRef] [Google Scholar]
  48. van Haarlem, M. P., Wise, M. W., Gunst, A. W., et al. 2013, A&A, 556, A2 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  49. Vaswani, A., Shazeer, N., Parmar, N., et al. 2017, Adv. Neural Inform. Process. Syst., 30 [Google Scholar]
  50. Walter, F., Decarli, R., Aravena, M., et al. 2016, ApJ, 833, 67 [Google Scholar]
  51. Westmeier, T., Kitaeff, S., Pallot, D., et al. 2021, MNRAS, 506, 3962 [NASA ADS] [CrossRef] [Google Scholar]

4

The produced data set is available by request.

5

The technical details are given in Appendix C.

6

The sky model x is of size N × M = 512×512. For the simplicity of notations we use the vectorized representation N·M.

7

The Python implementation of the proposed framework is publicly available at https://github.com/taranO/ml-based_source_localization_from_uv-plane

8

As it can be seen from top-right Fig. 6, the dynamic range of the magnitude is smaller than the dynamic range of the phase component. In this respect, we try to preserve this deviation and use the smaller normalization factor for the magnitude compared to the real and imaginary components.

9

The architecture details are given in Table D.1.

10

The architecture details are given in Table D.2.

11

The architecture details of the used auto-encoder are given in Table D.3.

12

More details are given in Appendix D.

13

The used PyBDSF parameters are thresh_pix = 7 and thresh_isl = 5. In the proposed framework area_lim parameter was set to 125 in case of Re & Im at stage 1 and to 200 at stage 2. In Mag & Phase, the same parameter was set to 300 and 310 at stage 1 and stage 2 correspondingly.

14

We set PyBDSF parameter rms to 4.2450E-05 for the CLEAN data and to 4.275E–05 for the dirty case. In the proposed framework area_lim parameter was set to 75 in case of Re & Im at stage 1 and to 240 at stage 2. In Mag & Phase, the same parameter was set to 200 and 340 at stage 1 and stage 2 correspondingly.

15

The size of the beam used in the simulation is about 0.82″.

16

In Fig. 11, we can see significant deviation from the average for Re & Im stage 2 in comparison to Re & Im stage 1. This explains why the performance of stage 2 is better than that of stage 1 in Fig. 14.

17

The inference of the proposed framework might be efficiently run at GPU. However, taking into account that the traditional methods are usually run on CPU and that not all users have access to GPU, we performed the comparison with respect to CPU to have a fair comparison and to show that the proposed framework is easily runnable on CPU.

All Tables

Table 1

Summary of the data simulation parameters using CASA.

Table 2

Number of sky models with different numbers of sources in the test subsets.

Table 3

Execution time for the source localisation per sky model (noisy data).

Table D.1

Architecture of the gθ model for ȳ ∊ ℝK represented by the real and imaginary components.

Table D.2

Architecture of the gθ model for ȳ ∊ ℝK represented by the magnitude and phase components.

Table D.3

Architecture of the qϕ model.

Table E.1

Performance (in%) of the state-of-the-art source detectors on the CLEAN images in the noise-free case.

Table E.2

Performance (in%) of the state-of-the-art source detectors on the CLEAN images in the noisy case, averaged over all S/Ns.

All Figures

thumbnail Fig. 1

Examples of simulated data used in this study. Top-left: true sky model. Top-middle: noise-free model. Top-right: noisy. Bottom-left: true sky model. Bottom-middle: noise-free. Bottom right: noisy. For both noise-free and noisy cases, the CLEAN representations are visualized. The sources are highlighted as red circles.

In the text
thumbnail Fig. 2

Example of closely located sources of different intensity. For better visibility, the cropped and zoomed parts of the true sky models are shown.

In the text
thumbnail Fig. 3

Schematic representation of the traditional pipeline.

In the text
thumbnail Fig. 4

Schematic representation of the proposed framework.

In the text
thumbnail Fig. 5

Detailed architecture of the gθ model, where ȳ denotes the normalized input data . Left panel: ȳ ∊ ℝK is represented by the real and imaginary components. Right panel: ӯ ∊ ℝK is represented by the magnitude and phase components.

In the text
thumbnail Fig. 6

Examples of input data. Top-left: real and imaginary components in the noise-free case. Top-right: magnitude and phase components in the noise-free case. Bottom-left: real and imaginary components in the noisy case. Bottom-right: magnitude and phase components in the noisy case.

In the text
thumbnail Fig. 7

Example of the proposed model’s outputs. Left to right: example of the true binary source map m ∊ {0,1}N×M. Proposed framework estimations, m^st.1N×M${{{\bf{\hat m}}}_{st.1}} \in {^{N \times M}}$ and m^st.2N×M${{{\bf{\hat m}}}_{st.2}} \in {^{N \times M}}$ based on the real and imaginary input data representation (for the magnitude and phase representation the results look similar). Final estimation: m^{ 0,1 }N×M${\bf{\hat m}} \in {\left\{ {0,1} \right\}^{N \times M}}$

In the text
thumbnail Fig. 8

Schematic representation of the acceptable distance of R between the true and predicted sources.

In the text
thumbnail Fig. 9

Purity and completeness of the methods under investigation in the noise-free case. Small black lines on the top of each bar correspond to the standard deviation from the average.

In the text
thumbnail Fig. 10

Difference between the noise-free (left) and noisy (right) prediction of the proposed Re & Im stage 1 for a simulated sky model.

In the text
thumbnail Fig. 11

Purity and completeness of the methods under investigation in the noisy case, averaged over all S/N values. The small black lines on the top of each bar correspond to the standard deviation from the average.

In the text
thumbnail Fig. 12

Dependence of completeness on S/N of the sources.

In the text
thumbnail Fig. 13

Dependence on the efficiency of PyBDSF and the proposed Re & Im framework on the number of sources in the sky model. Left: one source. Middle: three sources. Right: five sources.

In the text
thumbnail Fig. 14

Dependence on α, where α=Rbeam size$\alpha = {R \over {{\rm{beam}}\,{\rm{size}}}}$ and R is the distance between the true and predicted sources, in the noisy case, averaged over all S/N values.

In the text
thumbnail Fig. 15

Mean distance (the y/-axis) between the true and predicted sources in the noisy case. The coloured semi-transparent background shows the standard deviation from the mean value. The gray bars show the relative number of detected sources.

In the text
thumbnail Fig. 16

Dependence of completeness on normalized S/N compared to Béthermin et al. (2020).

In the text
thumbnail Fig. 17

Purity and completeness of PyBDSF on CLEAN data and the proposed Re & Im framework for the different noise levels, averaged over all S/Ns. From top left to bottom right: PWV = 0.472; PWV = 0.55; PWV = 0.658; PWV = 0.72; PWV = 0.913; PWV = 1.05; PWV = 1.262; PWV = 2.748; PWV = 3.9; PWV = 5.186; PWV = 0.

In the text
thumbnail Fig. A.1

Example of the total UV coverage for one of our ALMA 12-m array simulation pointing lasting 20 mins with 240 channels (top). The corresponding dirty beam is shown in the bottom panel.

In the text
thumbnail Fig. D.1

Dependence of the purity and completeness on the limit on the number of detected sources p.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.