Free Access
Issue
A&A
Volume 658, February 2022
Article Number A110
Number of page(s) 11
Section Numerical methods and codes
DOI https://doi.org/10.1051/0004-6361/202141874
Published online 07 February 2022

© ESO 2022

1. Introduction

The search for point sources (PS) as contaminants to the recovery of the cosmic microwave background (CMB) anisotropies at small angular scales has become more and more relevant through the years since the conception of the Wilkinson Microwave Anisotropy Probe (WMAP, Bennett et al. 2003) and Planck (Planck Collaboration IV 2020) missions. At millimetre wavelengths, most of the PS are dusty galaxies, mainly dusty star-forming galaxies among which there are the most intense stellar nurseries in the Universe, and blazars, that is active galactic nuclei (AGNs), whose jets are aligned in the line of sight of the satellite instrument. In this regime, they are one of the main contaminants to the recovery of the CMB anisotropies on small scales. The future CMB experiments, such as the Probe of Inflation and Cosmic Origins (PICO, Hanany et al. 2019), the CMB-S4 (Abazajian et al. 2019) and the Simons Observatory (SO, Ade et al. 2019), all of them with higher resolution than Planck, are designed to keep the PS contamination low. This can be achieved by measuring at frequencies with lower PS contribution, by taking the data in regions with robust extragalactic surveys available, and mostly by developing high-performance methods for PS detection.

Multi-frequency detection of PS is an important field of research because the Planck successors will be able to observe the sky at simultaneous wavelengths. The diffuse component separation is possible with multi-wavelength information, but the PS detection is not because they can have different physical properties, that is, each individual galaxy acting like a PS has its own unique spectral behaviour. Therefore, the PS detection is generally a task for single-frequency methods, and the catalogues of extragalactic sources are extracted from CMB maps one channel at a time. However, using this approach, valuable information that multi-wavelength experiments can offer is wasted. Thus, it is appropriate to develop accurate multi-frequency PS detection methods for future CMB experiments.

The first multi-frequency compact source detection methods made use of prior knowledge of the spatial profile of the sources. Moreover, they needed to know that sources should appear as compact objects in diffuse random fields, having the source located in a noisy map in order to help the detection. Some of those methods were wavelet techniques (Vielva et al. 2001, 2003; González-Nuevo et al. 2006; Sanz et al. 2006; López-Caniego et al. 2007), Bayesian approaches (Hobson & McLachlan 2003; Carvalho et al. 2009), linear filters (Sanz et al. 2001; Chiang et al. 2002; Herranz et al. 2002a,b; López-Caniego et al. 2004, 2005a,b), and matched filters (Tegmark & de Oliveira-Costa 1998; Barreiro et al. 2003; López-Caniego et al. 2006).

An evolution of these methods was based on combining simulated multi-wavelength maps to obtain a higher average in the signal-to-noise ratio of the sources. This approach was used in Naselsky et al. (2002). Similarly, Chen & Wright (2008) combined WMAP W and V bands, obtaining CMB-free maps that helped them to detect the fainter sources. Although these methods detected a higher number of sources, they failed to recover their flux density.

The matrix filters (MTXFs; Herranz & Sanz 2008) were introduced as an intermediate approach of the above mentioned methods. It looked for the detection of compact sources using their distinctive spatial behaviour and multi-wavelength information, but without the need to know their frequency dependence. After being applied to realistic microwave sky LFI Planck simulations (Herranz et al. 2009), MTXFs were used in Planck Collaboration XXVIII (2014) to validate the LFI channels of the Planck catalogue of compact sources (PCCS). After that, they were used in Planck Collaboration LIV (2018) to produce the Planck multi-frequency catalogue of non-thermal (i.e. synchrotron-dominated) sources (PCNT), which was the first multi-frequency catalogue of compact sources covering the nine Planck channels.

In the last years, machine learning techniques have represented a revolution in astrophysics and cosmology. One of the most popular machine learning models concerns neural networks. These are artificial intelligence techniques inspired by the human brain, and they can be trained to learn non-linear behaviours from data by supervised or unsupervised learning (Suárez Gómez et al. 2020, 2021; García Riesgo et al. 2019, 2020). These characteristics make neural networks a good candidate to obtain interesting results in cosmology. Some examples of recent applications of neural networks in this field are the identification of galaxy mergers (Pearson et al. 2019) and strong gravitational lenses (Petrillo et al. 2017; Hezaveh et al. 2017) in astronomical images, a galaxy classifier (Kim & Brunner 2017), a better estimation of cosmological constraints from weak lensing maps (Fluri et al. 2019), a real-time gravitational wave detector (George & Huerta 2018), a high-fidelity generation of weak lensing convergence maps (Mustafa et al. 2019), and cosmological structure formation simulations under different assumptions (Mathuriya et al. 2018; He et al. 2019; Perraudin et al. 2019; Giusarma et al. 2019).

Multi-layer perceptron (Juez et al. 2012) and convolutional neural networks (CNNs) (LeCun et al. 1989, 2015) have been successfully applied to image processing (and related fields) for modelling and forecasting (Graves et al. 2013; Giusti et al. 2013). An evolution of CNNs are the fully-convolutional neural networks (FCN) (Long et al. 2015; Dai et al. 2016), which are usually applied in image segmentation to classify a labelled object pixel-at-pixel by making use of different layers to learn image patterns (e.g., shapes, smoothness, and borders). In order to obtain relevant features of the input images, convolution and merging layers are generally paired in convolution and deconvolution steps. After that, the output (image or numerical) is obtained.

Fully-convolutional networks were recently used to detect PS in noisy microwave background simulations (Bonavera et al. 2021). The neural network, called PoSeIDoN, was trained at 217 GHz, and its performance was compared at 143, 217, and 353 GHz against the Mexican hat wavelet 2 (MHW2; González-Nuevo et al. 2006), a single-frequency filtering method that was used in the Planck experiment to detect PS at both LFI and HFI channels, creating the Planck catalogue of compact sources (Planck Collaboration XXVIII 2014) and the second Planck catalogue of compact sources (Planck Collaboration XXVI 2016). PoSeIDoN obtained similar results on completeness with respect to the filter but with a much lower number of spurious detections.

In this work, we propose the use of FCN as a multi-frequency generalisation of PoSeIDoN to develop the Multifrequency Point Source Image Detection Network (MultiPoSeIDoN) to detect compact sources in noisy background maps by using image segmentation.

The outline of the paper is the following. Section 2 covers how the simulated maps are generated. Section 3 describes our methodology. The results are explained in Sect. 4, and our conclusions are discussed in Sect. 5.

2. Simulations

In this work, realistic simulated maps of the microwave sky are used. These maps correspond to sky patches at the central channels of the Planck mission: 143, 217, and 353 GHz. The pixel size is 90 arcsec, which is a round number close to the 1.72 arcmin used in the Planck maps, corresponding to Nside = 2048 in the HEALPix all sky pixelisation schema (Górski et al. 2005). Balancing between the density of bright sources per patch and size, we consider that patches of 128 × 128 pixels are sufficient for our study.

The multi-frequency PS detection is studied with two different sets of simulations. The first one consists of realistic simulated maps whose PS contribution is from flat-spectrum sources, that is, objects that have the same spectral behaviour in the three channels. The other set of simulations have PS with different spectral behaviours, thus also including in this case sources that increase and decrease their emission with frequency.

In the first case, where only sources with a flat spectrum have been considered, radio galaxies, mainly radio quasars and BL lacertae objects (or simply BL Lacs), commonly known as blazars, are simulated at 217 GHz using the model by Tucci et al. (2011) and the software CORRSKY (González-Nuevo et al. 2005). Then, in order to produce the simulated catalogue at 143, 217, and 353 GHz, the flux densities are convolved with the corresponding instrumental FWHM (7.22, 4.90, and 4.92 arcmin at 143, 217, and 353 GHz, respectively (Planck Collaboration IV 2020). Although not substantially contributing to the statistical properties of the background, the infrared late-type galaxies (IRLT, mainly starburst and local spiral galaxies; Toffolatti et al. 1998) are also simulated in order to add their shot-noise contribution to the patches by adopting the source number counts by Cai et al. (2013), normalised to the later update by Negrello et al. (2013) and the software CORRSKY.

In order to obtain the set of simulations for the second case, the spectral behaviour of radio and IRLT sources is assumed to vary as

S = S 0 ( ν ν 0 ) α , $$ \begin{aligned} S = S_{0} \left( \frac{\nu }{\nu _{0}} \right)^{\alpha }, \end{aligned} $$(1)

where S is the flux density of the sources at each frequency ν, S0 is the flux density at the central channel with frequency ν0, that is, 217 GHz, and α is the spectral index.

The spectral index distribution for each population is estimated in Planck at each frequency (Planck Collaboration XXVI 2016). For frequencies below 217 GHz, the emission is mostly due to radio galaxies. Above such a frequency, most of the contribution comes from the late-type galaxies. At 217 GHz, the emission of both kind of sources is relevant to the total distribution. In our simulations, the spectral index is randomly chosen according to the Gaussian distributions with mean and standard deviation given in Planck Collaboration XXVI (2016).

To obtain realistic simulations at these frequencies, we also consider the cosmic infrared background (CIB, Puget et al. 1996; Hauser & Dwek 2001; Dole et al. 2006) emission due to high redshift infrared PS (massive proto-spheroidal galaxies in the process of forming most of their stellar mass, Granato et al. 2004; Lapi et al. 2006, 2011; Cai et al. 2013), which is dominant at a few arcminute resolution, resulting in a contaminant for the purpose of our work. To simulate them, the source number counts given by Cai et al. (2013), their angular power spectrum given by Lapi et al. (2011), and the software CORRSKY are used. Their spectral behaviour is simulated by assigning them the same spectral index as for the late-type galaxies. For all the source populations (radio, late-type, and CIB) and frequencies, we simulate them down to the same flux density limit of 3 μJy.

On larger angular scales, the contamination due to diffuse emission by our Galaxy and the CMB are also considered. Galactic emission varies significantly with frequency. Since our simulations correspond to the 143, 217, and 353 GHz Planck channels, only thermal dust emission (Finkbeiner et al. 1999) from our Galaxy is simulated. Both CMB and thermal dust emissions are introduced in the simulated maps by randomly selecting patches at the same position where the sources were simulated. On one hand, the CMB maps are the ones given at each frequency by the SEVEM method (Martínez-González et al. 2003; Leach et al. 2008; Fernández-Cobos et al. 2012). These maps are the inpainted version to fill the empty pixels due to the mask (both for the galactic plane and the PS) used during the component separation process (see Planck Collaboration IV 2020, and references therein). On the other hand, the thermal dust emission is added at each frequency by using the Planck Legacy Archive (PLA) simulations, which were produced using the Planck Sky Model software (Delabrouille et al. 2013).

The Planck maps are at Nside = 2048, which corresponds to a pixel size of 1.72 arcmin. The selected sky patches, which are randomly chosen to be negligible the probability of having two exact sky maps, are projected into flat patches of pixel size of 1.5 arcmin by using the gnomview function of HEALPix framework (Górski et al. 2005). Small-scale fluctuations are added to Galaxy emission by following the Miville-Deschênes et al. (2007) method in order to simulate the sky at Planck resolution. This method tries to reproduce the non-Gaussian behaviours of the interstellar emission by increasing its fluctuation level as a function of the local brightness (Leach et al. 2008).

The thermal Sunyaev-Zel’dovich effect produced by galaxy clusters is also considered. This effect is generally negligible in PS detection, but it is added for completeness. To simulate it, PLA simulations are also used at each frequency, extracting the patches at the same sky position as for the other components (Delabrouille et al. 2002, 2013). Moreover, instrumental noise is added to the simulations by considering white noise using the Planck values: 0.55, 0.78, and 2.56 μKCMB deg for 143, 217, and 353 GHz, respectively (Planck Collaboration IV 2020).

In our work, we considered all the contributions listed above as background ones, except for the PS. Overall, the background at 143 GHz is mainly the emission from the CMB, and it decreases while increasing the frequency. On the other hand, the contamination due to our Galaxy and proto-spheroidal galaxies increases at higher frequencies.

Examples of random simulated patches at the same location in the sky are shown in the first two columns of Fig. 1 for 143, 217, and 353 GHz (top, middle, and bottom panels, respectively) at b > 30° Galactic latitudes. The first column shows the background with the PS emission (i.e. the total images). The second column represents the PS-only map.

thumbnail Fig. 1.

From left to right, one simulation for the spectral case with the total and PS-only input validation maps, and the MTXFs and MultiPoSeIDoN PS outputs at 143, 217, and 353 GHz from top to bottom at b > 30° Galactic latitudes. The flux density values (in Jy) for each panel are shown in the colour bars.

3. Methodology

3.1. MultiPoSeIDoN

Mitchell and Goodfellow stated that a computer program is said to learn from experience with respect to a class of tasks and performance measures, if its performance improves with experience (Mitchell 1997; Goodfellow et al. 2016). The method by which a computer program learns is called machine learning. Neural networks are machine learning models inspired by the human brain with the goal of learning non-linear behaviours from the data. They are formed by connected layers of neurons, which are their basic computing units. Neurons have weights that adjust their value through a cost function on each step of training.

When the data have a known grid-like topology, the models used are called convolutional neural networks (CNN, LeCun et al. 1989, 2015). In this case, the weights correspond to kernel values, which are tensor-shaped arrays that model the connections between neurons. Every CNN is formed by convolutional blocks. Each of them consist of a layer that performs convolutions in parallel, followed by a set of linear activations, and by a pooling function, which aggregates information by grouping neighbouring pixels generally using their maximum or average values.

The fully-convolutional neural networks (FCN, Long et al. 2015), which aim to classify each pixel instead of the whole image to perform object segmentation, represent an evolution of this model. They make both learning and inference on the whole image at once through extracting the most relevant characteristics of the image by using convolutional blocks while making a prediction at each pixel by using deconvolutional blocks.

FCN, as any other neural network model, perform optimisation procedures to obtain parameters from data through minimising a loss function. To drive the loss function to a minimum, a gradient-based optimiser is used (Cauchy 1847). When the loss function is adjusted, an algorithm called backpropagation (Rumelhart et al. 1986) flows the information backward through the network, changing the weights of the neurons. At the same time, the optimiser learns by obtaining an unbiased estimate of the gradient on a small set of samples called minibatch (Kiefer & Wolfowitz 1952).

FCN are well-suited models for PS detection, as was concluded in Bonavera et al. (2021), since they make inferences through their convolutional blocks to detect PS in noisy maps, while they make predictions through their deconvolutional blocks to produce cleaned PS maps. With these output images, the catalogues can be created and studied statistically.

MultiPoSeIDoN is the FCN developed in this work to detect PS in noisy, multi-frequency background maps. It is a U-Net-based neural network (Ronneberger et al. 2015) with two different approaches: flat and spectral MultiPoSeIDoNs. The first one is trained with a set of 50 000 simulations of background and PS at 143, 217, and 353 GHz as inputs and a set of 50 000 simulations of PS-only at 217 GHz as labels. The second one is trained with a set of 50 000 simulations of background and PS and a set of 50 000 simulations of PS-only. In this case, both sets of simulations are at 143, 217, and 353 GHz as inputs and labels, respectively. Therefore, 200 000 and 300 000 images in total are used to train flat and spectral MultiPoSeIDoNs, respectively. Both training procedures are performed during 500 epochs (when an entire dataset is passed forward and backward through the neural network only once) by using a mean-squared error loss function (MSE, Hastie et al. 2001) and the adaptive gradient algorithm (AdaGrad, Duchi et al. 2011) to perform the learning with a rate of 0.05 on each minibatch of 32 samples. These hyperparameters were selected through a grid search.

The architectures for both flat and spectral MultiPoSeIDoNs are detailed as follows (top and bottom pannel in Fig. 2 respectively). First, the neural networks have a set of convolutional blocks: flat MultiPoSeIDoN have six convolutional blocks containing both convolutional and pooling layers with 8, 2, 4, 2, 2, and 2 kernels of sizes of 9, 9, 7, 7, 5, and 3, respectively. Its feature maps are 8, 16, 64, 128, 256, and 512, respectively. On the other hand, spectral MultiPoSeIDoN have the same number of layers, kernels, and kernel sizes, but, since we are using three label images instead of one, as for the flat case, we need to use threefold feature maps. In this case, they are 9, 18, 72, 144, 288, and 576. Both networks have a sub-sampling factor of 2. The padding type ‘Same’ (an additional layer that adds ‘space’ around the input data or the feature map, which helps to deal with possible loss in width and/or height dimensions in the feature maps after having applied the filters) is added in all the layers and the activation function is a leaky ReLU (Nair & Hinton 2010).

thumbnail Fig. 2.

Architecture of flat (top panel) and spectral (bottom panel) MultiPoSeIDoNs. The first one has a convolutional block, which produces eight feature maps. After that, the space dimensionality increases to 512 feature maps through five more convolutional blocks. The second one produces 9 and 576 feature maps in its first and last convolutional blocks, respectively. These layers are connected to deconvolutional ones, which decreases the space dimensionality to eight and nine feature maps in the last deconvolutional block for flat and spectral MultiPoSeIDoNs, respectively. Fine-grained features are added from each convolution to its corresponding deconvolution in both neural networks.

After that, the neural networks have a set of deconvolutional blocks: flat MultiPoSeIDoN convolutional blocks are connected to six inverse-convolutional (also called deconvolutional) plus pooling layers with 2, 2, 2, 4, 2, and 8 kernels of sizes of 3, 5, 7, 7, 9, and 9, respectively. Its feature maps are 256, 128, 64, 16, 8, and 1. On the other hand, the spectral MultiPoSeIDoN deconvolutional blocks are formed by the same number of layers, kernels, and kernel sizes, but its feature maps are 288, 144, 72, 18, 9, and 3, respectively. The sub-sampling factor is 2. The padding type Same is added in all the layers, and the activation function is a leaky ReLU. The feature maps resulting from the five last convolutions are added as fine-grained features to the results of the five first deconvolutions.

Both flat and spectral MultiPoSeIDoNs learn through their convolutional blocks that a PS is located at a given position in the background using the position and flux density information provided by the PS-only image, while their deconvolutional blocks perform a PS segmentation from the total input maps, resulting in a PS-only output image.

Both versions of MultiPoSeIDoN are trained with a large number of simulations to prevent underfitting, and they are tested during training using 5000 simulations isolated from the training dataset to prevent overfitting. However, since the objective is to predict a numerical flux density of the same type of object (i.e. a point in a map), overfitting is not a problem because the main goal is to deal with background in order to decrease the number of spurious sources (i.e. false positives) instead of detecting different objects in an image. However, learning curves of training and test errors have been used to prevent overfitting during training (Goodfellow et al. 2016). An example of a spectral MultiPoSeIDoN output patch (at 143, 217, and 353 GHz, from top to bottom) is shown in the last column of Fig. 1. Similar output images are obtained for the flat MultiPoSeIDoN case.

3.2. Matched matrix filters

The matched matrix filters (MTXFs, Herranz & Sanz 2008; Herranz et al. 2009) was introduced as a method to detect multi-frequency extragalactic PS using their distinctive spatial behaviour, multi-wavelength information, and the fact that their frequency dependence was not known a priori.

Let us consider a set of N two-dimensional images (channels) formed by PS and other foreground components mainly called generalised noise. The data model is defined by

D k ( x ) = s k ( x ) + n k ( x ) , $$ \begin{aligned} D_{k} (\mathbf x ) = s_{k} (\mathbf x ) + n_{k} (\mathbf x ), \end{aligned} $$(2)

where sk(x) is the term involving the PS and nk(x) is the one for the generalised noise. The subscript k = 1, ..., N is the index of each image. The term for the PS is defined as

s k ( x ) = A k τ k ( x ) $$ \begin{aligned} s_{k} (\mathbf x ) = A_{k} \tau _{k} (\mathbf x ) \end{aligned} $$(3)

where Ak is the unknown amplitude of the source in the kth channel and τk(x) is the known spatial profile of the source. On the other hand, for the generalised noise, the MTXFs assume a zero mean, i.e. ⟨nk(x)⟩ = 0, which can be characterised by its cross-power spectrum:

n k ( q ) n l ( q ) = P kl ( q ) δ 2 ( q q ) , $$ \begin{aligned} \langle n_{k} (\boldsymbol{q}) n_{l}^{*} (\boldsymbol{q}^{\prime }) \rangle = P_{k l} (\boldsymbol{q}) \delta ^{2} (\boldsymbol{q} - \boldsymbol{q}^{\prime }), \end{aligned} $$(4)

where P = (Pkl) is the cross-power spectrum matrix and ‘*’ denotes complex conjugation.

For accurate photometry in each channel, one has to produce N input maps and other N output maps. For a multi-wavelength approach, one can define a set of N × N filters Ψkl, which allows the input channels to help in the elaboration of the output maps. Then, the filtered images for a set of data Dl can be defined as

ω k ( x ) = l d x Ψ kl ( x x ) D l ( x ) = l d q e i q x Ψ kl ( q ) D l ( q ) , $$ \begin{aligned}&\omega _{k} (\boldsymbol{x}) = \sum _{l} \int \text{ d} {\boldsymbol{x}^{\prime }} \Psi _{k l} (\boldsymbol{x} - {\boldsymbol{x}^{\prime }}) D_{l} ({\boldsymbol{x}^{\prime }}) \nonumber \\&\qquad \ = \sum _{l} \int \text{ d} \boldsymbol{q} e^{-i \boldsymbol{q} \boldsymbol{x}} \Psi _{k l} (\boldsymbol{q}) D_{l} (\boldsymbol{q}), \end{aligned} $$(5)

where ωk is the sum of a set of linear filters of the data Dl. The root mean square (rms) of the filtered data is then estimated using the square root of the variance:

σ ω k 2 = l m d q Ψ kl ( q ) Ψ km P lm ( q ) . $$ \begin{aligned} \sigma _{\omega _{k}}^{2} = \sum _{l} \sum _{m} \int \text{ d} \boldsymbol{q} \Psi _{k l} (\boldsymbol{q}) \Psi _{k m}^{*} P_{l m} (\boldsymbol{q}). \end{aligned} $$(6)

The set of filters Ψ that minimise the variance σωk for all k channels, and for the same amplitudes Ak of the sources, as was derived in Herranz & Sanz (2008), can be obtained by matrix multiplication:

Ψ = F P 1 , $$ \begin{aligned} \mathbf \Psi ^{*} = \mathbf F \mathbf P ^{-1}, \end{aligned} $$(7)

where

Ψ = ( Ψ kl ) , F = ( F kl ) , P = ( P kl ) , λ = ( λ kl ) , H = ( H kl ) , $$ \begin{aligned}&\mathbf \Psi = (\Psi _{k l}), \mathbf F = (F_{k l}), \mathbf P = (P_{k l}), \nonumber \\&\boldsymbol{\lambda } = (\lambda _{k l}), \mathbf H = (H_{k l}), \end{aligned} $$(8)

are N × N matrices for any q, and

F kl = λ kl τ l , λ = H 1 , H kl = d q τ k ( q ) P kl 1 τ l ( q ) . $$ \begin{aligned}&F_{k l} = \lambda _{k l} \tau _{l}, \nonumber \\&\boldsymbol{\lambda } = \mathbf H ^{-1}, \nonumber \\&H_{k l} = \int \text{ d} \boldsymbol{q} \tau _{k} (\boldsymbol{q}) P_{k l}^{-1} \tau _{l} (\boldsymbol{q}). \end{aligned} $$(9)

When the filters (7) are applied to a set of N images, the filtered ones with values [ω1, ..., ωN] have the form

ω k = l d q Ψ kl ( q ) A l τ l ( q ) = A k . $$ \begin{aligned} \omega _{k} = \sum _{l} \int \text{ d} \boldsymbol{q} \Psi _{k l} (\boldsymbol{q}) A_{l} \tau _{l} (\boldsymbol{q}) = A_{k}. \end{aligned} $$(10)

To avoid the fact that the flux density of a PS in the lth channel could leak into the kth filtered image, the filters satisfy an orthonormality condition:

d q Ψ kl ( q ) τ l ( q ) = δ kl . $$ \begin{aligned} \int \text{ d} \boldsymbol{q} \Psi _{k l} (\boldsymbol{q}) \tau _{l} (\boldsymbol{q}) = \delta _{k l}. \end{aligned} $$(11)

Therefore, the filters are unbiased estimators of the flux density of the PS for all N channels. Moreover, to avoid the possibility of statistical correlations between the generalised noise in the lth and kth channels, the Hkl terms in Eq. (9) minimise the filtered rms σωk to lower values than the ones achieved by single-frequency matched filters (Herranz & Sanz 2008, Herranz et al. 2009).

Where the noise is uncorrelated through channels, the matrix of filters are diagonal matrices with elements:

Ψ kl ( q ) = δ kl τ k ( q ) / P k ( q ) d ( q ) τ k 2 ( q ) / P k ( q ) , $$ \begin{aligned} \Psi _{k l}^{*} (\boldsymbol{q}) = \delta _{k l} \frac{\tau _{k} (\boldsymbol{q}) / P_{k} (\boldsymbol{q}) }{\int \text{ d} (\boldsymbol{q}) \tau _{k}^{2} (\boldsymbol{q}) / P_{k} (\boldsymbol{q})}, \end{aligned} $$(12)

that is, the non-zero elements of the matrix filters are the complex conjugates of the matched filters of each channel. When the PS have symmetric profiles or statistically homogeneous and isotropic noise, the MTXFs show the same performance as the single-frequency matched filter, which is formed by real values. Since for real foreground-contaminated data, the generalised noise is neither homogeneous nor isotropic and is further correlated between channels, MTXFs are one of the best-suited methods to detecting multi-frequency PS.

With all these assumptions, one can conclude that MTXFs detect PS by removing the generalised noise by filtering in the scale domain of the sources and by cleaning out large-scale structures localised in neighbouring channels. An example of MTXF’s output patch (at 143, 217, and 353 GHz, from top to bottom) is shown in the third column of Fig. 1.

4. Results

Once we obtained the outputs of MultiPoSeIDoN and the MTXFs on the same validation simulations (a set of 5000 simulations isolated from train and testing datasets), we created the input and output (with the detections) catalogues by searching the local maxima of the sources, after introducing a threshold in flux density of 60 mJy for the three models, and also a safer 4σ limit for the MTXFs, as the one used in Planck catalogues. We also considered a detection when sources are separated by a two-pixel minimum distance. Border effects are one of the main problems of filtering techniques such as the MTXFs, because the detection algorithm could consider some artefacts near the patch border as spurious PS detections. As was concluded in Bonavera et al. (2021), neural networks does not have to deal with border effect issues, so the whole patch can be included in the analysis. However, for the MTXFs, some pixels in the patch border must be avoided. In our case, we removed five pixels both in width and height. To asses our results, we performed a statistical analysis of the catalogues resulting from the outputs of both flat and spectral MultiPoSeIDoNs, as well as for the MTXFs. The statistical quantities analysed are the completeness, the percentage of spurious sources, and the flux density comparison between the input and the recovered values (López-Caniego et al. 2007; Planck Collaboration VII 2011, 2014, 2016, 2020; Hopkins et al. 2015). Completeness is the ratio between the number of recovered true sources and the total number of input sources over a given flux limit. It is defined by the relation

C ( > S 0 ) = N true detected ( > S 0 ) N input ( > S 0 ) , $$ \begin{aligned} C (> S_{0}) = \frac{N_{\rm true detected} (> S_{0})}{N_{\rm input} (> S_{0})}, \end{aligned} $$(13)

where S0 is the input flux density, Ntrue detected(> S0) is the number of true detected sources by each method, and Ninput(> S0) is the number of sources in the PS-only input catalogue. The detected sources that do not have a counterpart in the input catalogue (i.e. false positives) are called spurious sources. Their number can be estimated using

R ( S 0 ) = N spurious ( > S 0 ) N input ( > S 0 ) = N output ( > S 0 ) N true detected ( > S 0 ) N input ( > S 0 ) , $$ \begin{aligned}&R (S_{0}) = \frac{N_{\rm spurious} (> S_{0})}{N_{\rm input} (> S_{0})} \nonumber \\&\qquad \ \ \ = \frac{N_{\rm output} (> S_{0}) - N_{\rm true detected} (> S_{0})}{N_{\rm input} (> S_{0})}, \end{aligned} $$(14)

where S0 is the input flux density, Noutput(> S0) is the number of detected sources after having used each method, Ntrue detected(> S0) is the number of detected sources which are in both the input and detection catalogues (i.e. true detections), and Ninput(> S0) is the number of sources in the PS-only input catalogue.

4.1. Completeness and spurious sources

The completeness (top sub-panel) and the percentage of spurious sources (bottom sub-panel) with respect to the input ones are shown on the left column of Fig. 3. For the MTXFs (cyan solid line), we obtain the expected results: using a detection limit of 60 mJy in flux density, the MTXFs provide good completeness results for the three Planck channels, reaching the 90% completeness level at 84, 79, and 123 mJy for 143, 217, and 353 GHz, respectively. We also used a 4σ limit (orange solid line) to reduce the number of spurious detections, reaching the 90% completeness level at 113, 92, and 398 GHz.

thumbnail Fig. 3.

Validation results of completeness (left column, top sub-panel), percentage of spurious sources (left column, bottom sub-panel), and flux density comparison (right column) between the input sources and the recovered ones by flat (blue series) and spectral (red series) MultiPoSeIDoNs, the MTXFs with a detection limit of 60 mJy (cyan series), and with a 4σ detection limit (orange series). The frequencies are 143 (top panels), 217 (middle panels), and 353 GHz (bottom panels) for a 30° Galactic cut. The dotted grey vertical line is the 90% completeness flux density limit (177, 152, and 304 mJy for 143, 217, and 353 GHz, respectively) for the second Planck catalogue of compact sources (Planck Collaboration XXVI 2016).

The percentage values of spurious sources are sensitive to detection limits in flux density; more than 20% of the total detected sources are spurious bellow 180 mJy at 143, and this behaviour increases with frequency, reaching the same percentage value below ∼400 and 1200 mJy for 217 and 353 GHz, respectively. The application of a 4σ detection limit for the flux density does not improve these results much.

Flat (blue solid line) and spectral (red solid line) MultiPoSeIDoNs have a relatively similar completeness results at 143 and 217 GHz, reaching the 90% completeness level at 58 mJy for a flat MultiPoSeIDoN and 79 and 71 mJy for a spectral MultiPoSeIDoN at 143 and 217 GHz, respectively. For 353 GHz, the completeness results of MultiPoSeIDoN are much better than for the MTXFs, reaching the 90% completeness level at 60 mJy. The clear advantage of MultiPoSeIDoN with respect to the MTXFs is shown in the percentage of spurious sources panel: both neural networks reach percentages of spurious sources of 20% below 100 mJy. Moreover, above the 90% of completeness level, they detect sources with a percentage of spurious sources between 0 and 10%. This behaviour is similar to that found in Bonavera et al. (2021), when a single-frequency neural network called PoSeIDoN was compared in completeness and percentage of spurious sources with the Mexican hat wavelet 2 (MHW2 González-Nuevo et al. 2006). The spurious PS issue is mainly due to a strongly contaminant background when the emission from our Galaxy and the CIB is higher. Therefore, obtaining lower values of spurious sources implies that MultiPoSeIDoN detects PS with higher accuracy than the MTXFs.

At 217 GHz, there is a peak in both flat and spectral cases, showing a high number of spurious sources above 1000 mJy up to about 4000 mJy. These peaks correspond to sources located at the borders of the map. This small issue is due to the catalogue production on the MultiPoSeIDoN outputs rather than the MultiPoSeIDoN detection capability.

By comparing the two versions of MultiPoSeIDoN, the flat one performs slightly better than the spectral one based on the results of completeness and reliability. This is explained by considering that spectral MultiPoSeIDoN have to deal with the spectral behaviour of PS, challenging their detection, as in a realistic case.

4.2. Photometry

In Fig. 3 (right column), we compare the recovered and input flux density of the sources. On one hand, the MTXFs correctly recover most of the flux densities above 100 mJy at 143 and 217, respectively, and above 150 mJy at 353 GHz within a 10% relative error. However, below those flux density values, the filters perform with a higher relative error, mainly due to very faint PS that are hardly detectable and could be located near a CMB fluctuation or a region with a high contamination from our Galaxy. This effect is called the Eddington bias (Eddington 1913), which is more visible at 353 GHz, when the contamination from the background is higher.

On the other hand, both flat (blue dots) and spectral (red dots) MultiPoSeIDoNs recover the flux density of the sources quite well, especially the brighter ones (flux densities above 100 mJy), the flat version being slightly better than the spectral one, as explained in the previous section. However, the results are still very good even at the fainter flux densities (already at 100 mJy for all the frequencies) with relative errors such as 5% for the flat case and between 7% and 10% for the spectral case.

At 217 GHz, a double diagonal can be seen above 300 mJy with a relative error of ±5%. This issue is due to a bad fitting of the fainter IRLT sources by the network. At this frequency, either IRLT and radio galaxies are in the training catalogue, as was explained before. More particularly, in this regime, the IRLT population is generally formed by fainter sources. This bad fitting is due to a not sufficiently accurate assigned spectral index value for the fainter IRLT sources, implying a worse performance recovering the flux density of the sources for the neural network at this frequency and flux density regime. However, MultiPoSeIDoN is removing part of the background that is increasing the flux density of those sources, which is the main objective of the network, and the detection is not affected at all by this issue attending on the results of completeness and spurious detections.

4.3. Multi-frequency versus single-frequency point source detection with fully convolutional networks

PoSeIDoN is a fully convolutional neural network designed to detect PS in single-frequency noisy background maps. Its performance was evaluated in Bonavera et al. (2021) by training it at 217 GHz with 50 000 realistic simulated images of PS plus contaminants such as CMB, CIB and thermal dust emissions, the thermal Sunyaev-Zel’dovich effect, and instrumental noise. It learned to detect PS with the help of 50 000 label images formed by PS only (radio and IRLT galaxies). Its performance was compared to the Mexican hat wavelet 2 filter (MWH2, González-Nuevo et al. 2006), one of the most popular single-frequency PS detection methods, which was used in the Planck experiment to validate both LFI and HFI channels and to create both PCCS and PCCS2 (Planck Collaboration XXVIII 2014; Planck Collaboration XXVI 2016). The results have shown that PoSeIDoN has a similar completeness results to the filter, but with much lower number of spurious detections, not only at the frequency at which it was trained, but also at 143 GHz and 353 GHz. Therefore, they concluded that PoSeIDoN is a more accurate and robust method.

The three statistical quantities explained above are used to compare the performance between PoSeIDoN and MultiPoSeIDoN. Both neural networks are validated with the same 5000 simulations at 217 GHz. The detection limit of 60 mJy in flux density and the two-pixel minimum distance between the sources to create the detection catalogues are also the same for both models.

As is shown in Fig. 4, the completeness results (left column, top sub-panel) show that both flat and spectral MultiPoSeIDoNs (blue and red solid lines respectively) show better performance than the single-frequency neural network (green solid line), reaching the 90% completeness level at 58 and 71 mJy, respectively, while PoSeIDoN reaches it at ∼300 mJy. On the other hand, the percentage of spurious sources (left column, bottom sub-panel) shows that the detections of MultiPoSeIDoN are more accurate since 20% of the PoSeIDoN detections are spurious below 150 mJy, while for MultiPoSeIDoN the same percentage of spurious sources is obtained below 100 mJy.

thumbnail Fig. 4.

Validation results of completeness (left column, top sub-panel), percentage of spurious sources (left column, bottom sub-panel), and flux density comparison (right column) between the input sources and the recovered ones by Flat (blue series) and Spectral (red series) MultiPoSeIDoNs and PoSeIDoN (green series). The frequency is 217 GHz for a 30° Galactic cut. The dotted grey vertical line is the 90% completeness flux density limit (152 mJy) for the second Planck catalogue of compact sources (Planck Collaboration XXVI 2016).

The comparison between the flux density of input sources (from the validation set) and of detected sources is shown at the right column in Fig. 4 for the three neural networks. On one hand, PoSeIDoN (green dots) tends to underestimate the flux density of the recovered fainter sources (which was exactly the opposite behaviour to the MHW2, as is shown in Bonavera et al. 2021). On the other hand, both flat and spectral MultiPoSeIDoNs (blue and red dots, respectively) show much better performance in recovering the flux density of the fainter PS, with only a relative error between 5% and 10% below 100 mJy, while PoSeIDoN has an up to 80% relative error below the same flux density limit. For the more intense sources, the three neural networks recover the flux density of the detections quite well.

5. Conclusions

In this work, we successfully applied a modern approach for the multi-frequency detection of point sources (PS) based on fully-convolutional neural networks trained with realistic simulations of images of the microwave sky. The patches of the images have contributions from PS (radio and IR late-type galaxies), the emission from the CMB and from the thermal dust of our Galaxy, the contamination due to massive proto-spheroidal galaxies (CIB), the thermal Sunyaev-Zel’dovich effect, and instrumental noise using Planck values. The CMB maps are the ones from the SEVEM method, and the Galaxy and Sunyaev-Zel’dovich contributions are those provided by the PLA. The frequencies were 143, 217, and 353 GHz at b > 30° Galactic latitudes.

A total of 50 000 simulations of background plus PS and 50 000 PS-only simulations were used to train two neural networks: flat MultiPoSeIDoN is trained by assuming sources with a flat spectrum (i.e. the sources had the same emission in all the channels), and spectral MultiPoSeIDoN uses simulations considering sources with a spectral behaviour. In that case, the emission from radio galaxies is higher at 143 GHz, whereas at 353 GHz IR late-type galaxies have more impact on the maps. The spectral behaviour is simulated by using the spectral index of each population after fitting their gaussian distributions plotted in PCCS2 (Planck Collaboration XXVI 2016).

After the training, 5000 simulations (isolated from the training dataset) are used to validate both flat and spectral MultiPoSeIDoNs and our comparison method, the matrix filters (MTXFs). Then, they are compared using their completeness, percentage of spurious sources, and flux density estimation values. Applying a detection limit of 60 mJy, MultiPoSeIDoN performs better than the MTXFs, especially when dealing with spurious sources. It reaches the 90% completeness level at 58 mJy for the flat case and 79, 71, and 60 mJy for the spectral case at 143, 217, and 353 GHz, respectively. The filter reaches the 90% completeness level at 84, 79, and 123 mJy. Of these detections, the percentage of spurious ones detected by the neural networks is 20% below a flux density of 100 mJy, decreasing it to values between 0 and 10% above that flux density value. For the filter, the percentage of spurious detections is higher than 20% below 180 mJy at 143 GHz. For the other frequencies, more than 20% of detections are spurious below ∼400 and 1200 mJy. Using the 4σ limit does not help to improve these results.

Based on these results, MultiPoSeIDoN has a similar completeness to the MTXFs at 143 and 217 GHz, but it is better at 353 GHz. For all those detections, the ones from the neural network are more reliable, especially for the sources with higher intensity, although we used a safer 4σ detection limit. As expected, the performance of the matrix filter is worse at higher frequencies, when the contamination from our Galaxy and the CIB is higher. However, the neural network demonstrates a rather equal performance at these higher frequencies. Based on that behaviour, we can assume that the parameters that model the spectral behaviour of the foreground components are successfully learned by MultiPoSeIDoN. Therefore, we anticipate that the higher contamination from the foregrounds would only weakly affect the performance of the neural network at higher frequencies than 353 GHz, or in regions inside the Galactic mask, where the contamination from our Galaxy has limited the Planck Collaboration when it comes to creating a reliable catalogue of compact sources in those regions. However, those two possibilities are outside the scope of this work.

The results on photometry show that both flat and spectral MultiPoSeIDoNs perform well in recovering the flux density of the sources, especially the brighter ones. On the other hand, the MTXFs are affected by the Eddington Bias at higher flux densities than the neural networks.

Another advantage of MultiPoSeIDoN with respect to the MTXFs is that we do not need to subtract pixels from the patch borders to avoid the border effect issue of the filtering methods. Therefore, the whole patch was considered in the analysis, allowing MultiPoSeIDoN to detect sources located at the patch border.

The comparison between both neural networks shows that flat MultiPoSeIDoN performs slightly better than the spectral case for the three statistical quantities. However, spectral MultiPoSeIDoN successfully obtains results considering that it is a more general and realistic method.

MultiPoSeIDoN can be considered as an evolution of its single-frequency relative, PoSeIDoN, because the performance of the multi-frequency approach is better than the one in the single-frequency case, especially regarding the photometry results: MultiPoSeIDoN recovers the flux density of the fainter sources with a much lower relative error (a mean difference of ∼70% with respect the results of PoSeIDoN). There are two main reasons for that: the first one is the increasing of the training information. The flat case was trained with twice the amount of information for PoSeIDoN, and the spectral one had three times the information. The second reason is that MultiPoSeIDoN learned the different correlations between the elements in the simulations due to their spectral behaviours.

As a future perspective, MultiPoSeIDoN could be trained for other CMB experiments such as the Q-U-I JOint Tenerife experiment (QUIJOTE, Rubiño-Martín et al. 2012), LiteBird, or PICO. It could be used to study CMB P-polarization modes to compare its performance against a recent Bayesian method, which uses images of realistic simulations with P-polarized compact sources (Herranz et al. 2021). Another interesting study could be to use MultiPoSeIDoN in multi-frequency component separation, training it to segmentate the CMB from the other foreground components.

Acknowledgments

We warmly thank the anonymous referee for the very useful comments on the original manuscript. J.M.C., J.G.N., L.B., M.M.C. and D.C. acknowledge financial support from the PGC 2018 project PGC2018-101948-B-I00 (MICINN, FEDER). DH acknowledges the Spanish MINECO and the Spanish Ministerio de Ciencia, Innovación y Universidades for partial financial support under project PGC2018-101814-B-I00. M.M.C. acknowledges PAPI-20-PF-23 (Universidad de Oviedo). J.D.C.J., M.L.S., S.L.S.G., J.D.S. and F.S.L. acknowledge financial support from the I+D 2017 project AYA2017-89121-P and support from the European Union’s Horizon 2020 research and innovation programme under the H2020-INFRAIA-2018-2020 grant agreement No 210489629. This research has made use of the python packages ipython (Pérez & Granger 2007), matplotlib (Hunter 2007), TensorFlow (Abadi et al. 2015), Numpy (Oliphant 2006) and Scipy (Jones et al. 2001), also the HEALPix (Górski et al. 2005) and healpy (Zonca et al. 2019) packages.

References

  1. Abadi, M., Agarwal, A., Barham, P., et al. 2015, TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, software available from http://tensorflow.org/ [Google Scholar]
  2. Abazajian, K., Addison, G., Adshead, P., et al. 2019, CMB-S4 Decadal Survey APC White Paper [CrossRef] [Google Scholar]
  3. Ade, P., Aguirre, J., Ahmed, Z., et al. 2019, JCAP, 2019, 056 [Google Scholar]
  4. Barreiro, R. B., Sanz, J. L., Herranz, D., & Martínez-González, E. 2003, MNRAS, 342, 119 [NASA ADS] [CrossRef] [Google Scholar]
  5. Bennett, C. L., Bay, M., Halpern, M., et al. 2003, ApJ, 583, 1 [NASA ADS] [CrossRef] [Google Scholar]
  6. Bonavera, L., Suarez Gomez, S. L., González-Nuevo, J., et al. 2021, A&A, 648, A50 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  7. Cai, Z.-Y., Lapi, A., Xia, J.-Q., et al. 2013, ApJ, 768, 21 [NASA ADS] [CrossRef] [Google Scholar]
  8. Carvalho, P., Rocha, G., & Hobson, M. P. 2009, MNRAS, 393, 681 [NASA ADS] [CrossRef] [Google Scholar]
  9. Cauchy, A. 1847, C.R. Acad. Sci. Paris, 25, 536 [Google Scholar]
  10. Chen, X., & Wright, E. L. 2008, ApJ, 681, 747 [NASA ADS] [CrossRef] [Google Scholar]
  11. Chiang, L. Y., Jørgensen, H. E., Naselsky, I. P., et al. 2002, MNRAS, 335, 1054 [NASA ADS] [CrossRef] [Google Scholar]
  12. Dai, J., Li, Y., He, K., & Sun, J. 2016, in Advances in Neural Information Processing Systems, 379 [Google Scholar]
  13. Delabrouille, J., Melin, J. B., & Bartlett, J. G. 2002, in Simulations of Sunyaev-Zel’dovich Maps and Their Applications, eds. L. W. Chen, C. P. Ma, K. W. Ng, & U. L. Pen, ASP Conf. Ser., 257, 81 [Google Scholar]
  14. Delabrouille, J., Betoule, M., Melin, J. B., et al. 2013, A&A, 553, A96 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  15. Dole, H., Lagache, G., Puget, J.-L., et al. 2006, A&A, 451, 417 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  16. Duchi, J., Hazan, E., & Singer, Y. 2011, J. Mach. Learn. Res., 12, 2121 [Google Scholar]
  17. Eddington, A. S. 1913, MNRAS, 73, 359 [Google Scholar]
  18. Fernández-Cobos, R., Vielva, P., Barreiro, R. B., & Martínez-González, E. 2012, MNRAS, 420, 2162 [Google Scholar]
  19. Finkbeiner, D. P., Davis, M., & Schlegel, D. J. 1999, ApJ, 524, 867 [NASA ADS] [CrossRef] [Google Scholar]
  20. Fluri, J., Kacprzak, T., Lucchi, A., et al. 2019, Phys. Rev. D, 100, 063514 [Google Scholar]
  21. García Riesgo, F., Suárez Gómez, S. L., Sánchez Lasheras, F., et al. 2019, in Hybrid Artificial Intelligent Systems, eds. H. Pérez García, L. Sánchez González, M. Castejón Limas, et al. (Cham: Springer International Publishing), 335 [CrossRef] [Google Scholar]
  22. George, D., & Huerta, E. A. 2018, Phys. Lett. B, 778, 64 [NASA ADS] [CrossRef] [Google Scholar]
  23. Giusarma, E., Reyes Hurtado, M., Villaescusa-Navarro, F., et al. 2019, ArXiv e-prints [arXiv:1910.04255] [Google Scholar]
  24. Giusti, A., Cireşan, D. C., Masci, J., Gambardella, L. M., & Schmidhuber, J. 2013, 2013 IEEE International Conference on Image Processing (IEEE), 4034 [Google Scholar]
  25. González-Nuevo, J., Toffolatti, L., & Argüeso, F. 2005, ApJ, 621, 1 [Google Scholar]
  26. González-Nuevo, J., Argüeso, F., López-Caniego, M., et al. 2006, MNRAS, 369, 1603 [Google Scholar]
  27. Goodfellow, I. J., Bengio, Y., & Courville, A. 2016, Deep Learning (Cambridge, MA, USA: MIT Press), http://www.deeplearningbook.org [Google Scholar]
  28. Górski, K. M., Hivon, E., Banday, A. J., et al. 2005, ApJ, 622, 759 [Google Scholar]
  29. Granato, G. L., De Zotti, G., Silva, L., Bressan, A., & Danese, L. 2004, ApJ, 600, 580 [Google Scholar]
  30. Graves, A., Mohamed, A. R., & Hinton, G. 2013, in 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (IEEE), 6645 [Google Scholar]
  31. Hanany, S., Alvarez, M., Artis, E., et al. 2019, ArXiv e-prints [arXiv:1902.10541] [Google Scholar]
  32. Hastie, T., Tibshirani, R., & Friedman, J. 2001, The Elements of Statistical Learning, Springer Series in Statistics (New York, NY, USA: Springer, New York Inc.) [CrossRef] [Google Scholar]
  33. Hauser, M. G., & Dwek, E. 2001, ARA&A, 39, 249 [Google Scholar]
  34. He, S., Li, Y., Feng, Y., et al. 2019, Proc. Nat. Acad. Sci., 116, 13825 [Google Scholar]
  35. Herranz, D., & Sanz, J. L. 2008, IEEE J. Sel. Top. Signal Process., 2, 727 [CrossRef] [Google Scholar]
  36. Herranz, D., Gallegos, J., Sanz, J. L., & Martínez-González, E. 2002a, MNRAS, 334, 533 [NASA ADS] [CrossRef] [Google Scholar]
  37. Herranz, D., Sanz, J. L., Barreiro, R. B., & Martínez-González, E. 2002b, ApJ, 580, 610 [Google Scholar]
  38. Herranz, D., López-Caniego, M., Sanz, J. L., & González-Nuevo, J. 2009, MNRAS, 394, 510 [NASA ADS] [CrossRef] [Google Scholar]
  39. Herranz, D., Argüeso, F., Toffolatti, L., Manjón-García, A., & López-Caniego, M. 2021, A&A, 651, A24 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  40. Hezaveh, Y. D., Perreault Levasseur, L., & Marshall, P. J. 2017, Nature, 548, 555 [Google Scholar]
  41. Hobson, M. P., & McLachlan, C. 2003, MNRAS, 338, 765 [NASA ADS] [CrossRef] [Google Scholar]
  42. Hopkins, A. M., Whiting, M. T., Seymour, N., et al. 2015, PASA, 32 [CrossRef] [Google Scholar]
  43. Hunter, J. D. 2007, Comput. Sci. Eng., 9, 90 [NASA ADS] [CrossRef] [Google Scholar]
  44. Jones, E., Oliphant, T., Peterson, P., et al. 2001, SciPy: Open source scientific tools for Python [Google Scholar]
  45. Juez, F. J., Lasheras, F. S., Roqueñí, N., & Osborn, J. 2012, Sensors, 12, 8895 [Google Scholar]
  46. Kiefer, J., & Wolfowitz, J. 1952, Ann. Math. Stat., 23, 462 [CrossRef] [Google Scholar]
  47. Kim, E. J., & Brunner, R. J. 2017, MNRAS, 464, 4463 [Google Scholar]
  48. Lapi, A., Shankar, F., Mao, J., et al. 2006, ApJ, 650, 42 [Google Scholar]
  49. Lapi, A., González-Nuevo, J., Fan, L., et al. 2011, ApJ, 742, 24 [NASA ADS] [CrossRef] [Google Scholar]
  50. Leach, S. M., Cardoso, J. F., Baccigalupi, C., et al. 2008, A&A, 491, 597 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  51. LeCun, Y., Boser, B., Denker, J. S., et al. 1989, Neural Comput., 1, 541 [NASA ADS] [CrossRef] [Google Scholar]
  52. LeCun, Y., Bengio, Y., & Hinton, G. 2015, Nature, 521, 436 [Google Scholar]
  53. Long, J., Shelhamer, E., & Darrell, T. 2015, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 3431 [Google Scholar]
  54. López-Caniego, M., Herranz, D., Barreiro, R. B., & Sanz, J. L. 2004, in Computational Imaging II, eds. C. A. Bouman, & E. L. Miller, SPIE Conf. Ser., 5299, 145 [CrossRef] [Google Scholar]
  55. López-Caniego, M., Herranz, D., Barreiro, R. B., & Sanz, J. L. 2005a, MNRAS, 359, 993 [CrossRef] [Google Scholar]
  56. López-Caniego, M., Herranz, D., Sanz, J. L., & Barreiro, R. B. 2005b, EURASIP J. Appl. Signal Process., 2005, 985049 [Google Scholar]
  57. López-Caniego, M., Herranz, D., González-Nuevo, J., et al. 2006, MNRAS, 370, 2047 [Google Scholar]
  58. López-Caniego, M., González-Nuevo, J., Herranz, D., et al. 2007, ApJS, 170, 108 [Google Scholar]
  59. Martínez-González, E., Diego, J. M., Vielva, P., & Silk, J. 2003, MNRAS, 345, 1101 [Google Scholar]
  60. Mathuriya, A., Bard, D., Mendygral, P., et al. 2018, ArXiv e-prints [arXiv:1808.04728] [Google Scholar]
  61. Mitchell, T. M. 1997, Machine Learning (New York: McGraw-Hill) [Google Scholar]
  62. Miville-Deschênes, M. A., Lagache, G., Boulanger, F., & Puget, J. L. 2007, A&A, 469, 595 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  63. Mustafa, M., Bard, D., Bhimji, W., et al. 2019, Comput. Astrophys. Cosmol., 6, 1 [Google Scholar]
  64. Nair, V., & Hinton, G. E. 2010, in Proceedings of the 27th International Conference on Machine Learning (ICML-10), 807 [Google Scholar]
  65. Naselsky, P., Novikov, D., & Silk, J. 2002, MNRAS, 335, 550 [NASA ADS] [CrossRef] [Google Scholar]
  66. Negrello, M., Clemens, M., Gonzalez-Nuevo, J., et al. 2013, MNRAS, 429, 1309 [Google Scholar]
  67. Oliphant, T. 2006, NumPy: A guide to NumPy (USA:Trelgol Publishing) [Google Scholar]
  68. Pearson, W. J., Wang, L., Alpaslan, M., et al. 2019, A&A, 631, A51 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  69. Pérez, F., & Granger, B. E. 2007, Comput. Sci. Eng., 9, 21 [Google Scholar]
  70. Perraudin, N., Srivastava, A., Lucchi, A., et al. 2019, Comput. Astrophys. Cosmol., 6, 5 [Google Scholar]
  71. Petrillo, C. E., Tortora, C., Chatterjee, S., et al. 2017, MNRAS, 472, 1129 [Google Scholar]
  72. Planck Collaboration VII. 2011, A&A, 536, A7 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  73. Planck Collaboration XXVIII. 2014, A&A, 571, A28 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  74. Planck Collaboration XXVI. 2016, A&A, 594, A26 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  75. Planck Collaboration LIV. 2018, A&A, 619, A94 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  76. Planck Collaboration I. 2020, A&A, 641, A1 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  77. Planck Collaboration IV. 2020, A&A, 641, A4 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  78. Puget, J. L., Abergel, A., Bernard, J. P., et al. 1996, A&A, 308, L5 [Google Scholar]
  79. Riesgo, F. G., Gómez, S. L. S., Rodríguez, J. D. S., et al. 2020, in Hybrid Artificial Intelligent Systems eds. E. A. de la Cal, J. R. Villar Flecha, H. Quintián, & E. Corchado (Cham: Springer International Publishing), 674 [Google Scholar]
  80. Ronneberger, O., Fischer, P., & Brox, T. 2015, Medical Image Computing and Computer-Assisted Intervention - MICCAI 2015 [Google Scholar]
  81. Rubiño-Martín, J. A., Rebolo, R., Aguiar, M., et al. 2012, in The QUIJOTE-CMB experiment: studying the polarisation of the galactic and cosmological microwave emissions, SPIE Conf. Ser., 8444, 84442Y [Google Scholar]
  82. Rumelhart, D. E., Hinton, G. E., & Williams, R. J. 1986, Nature, 323, 533 [Google Scholar]
  83. Sanz, J. L., Herranz, D., & Martínez-Gónzalez, E. 2001, ApJ, 552, 484 [NASA ADS] [CrossRef] [Google Scholar]
  84. Sanz, J. L., Herranz, D., Lopez-Caniego, M., & Argueso, F. 2006, ArXiv e-prints [arXiv:astro-ph/0609351] [Google Scholar]
  85. Suárez Gómez, S. L., González-Gutiérrez, C., Suárez, J. D., et al. 2020, Logic J. IGPL, 29, 180 [Google Scholar]
  86. Suárez Gómez, S. L., García Riesgo, F., González Gutiérrez, C., Rodríguez Ramos, L. F., & Santos, J. D. 2021, Mathematics, 9 [Google Scholar]
  87. Tegmark, M., & de Oliveira-Costa, A. 1998, ApJ, 500, L83 [Google Scholar]
  88. Toffolatti, L., Argueso Gomez, F., de Zotti, G., et al. 1998, MNRAS, 297, 117 [Google Scholar]
  89. Tucci, M., Toffolatti, L., de Zotti, G., & Martínez-González, E. 2011, A&A, 533, A57 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  90. Vielva, P., Barreiro, R., Hobson, M., et al. 2001, MNRAS, 328, 1 [NASA ADS] [CrossRef] [Google Scholar]
  91. Vielva, P., Martínez-González, E., Gallegos, J. E., Toffolatti, L., & Sanz, J. L. 2003, MNRAS, 344, 89 [Google Scholar]
  92. Zonca, A., Singer, L. P., Lenz, D., et al. 2019, J. Open Sour. Software, 4, 1298 [NASA ADS] [CrossRef] [Google Scholar]

All Figures

thumbnail Fig. 1.

From left to right, one simulation for the spectral case with the total and PS-only input validation maps, and the MTXFs and MultiPoSeIDoN PS outputs at 143, 217, and 353 GHz from top to bottom at b > 30° Galactic latitudes. The flux density values (in Jy) for each panel are shown in the colour bars.

In the text
thumbnail Fig. 2.

Architecture of flat (top panel) and spectral (bottom panel) MultiPoSeIDoNs. The first one has a convolutional block, which produces eight feature maps. After that, the space dimensionality increases to 512 feature maps through five more convolutional blocks. The second one produces 9 and 576 feature maps in its first and last convolutional blocks, respectively. These layers are connected to deconvolutional ones, which decreases the space dimensionality to eight and nine feature maps in the last deconvolutional block for flat and spectral MultiPoSeIDoNs, respectively. Fine-grained features are added from each convolution to its corresponding deconvolution in both neural networks.

In the text
thumbnail Fig. 3.

Validation results of completeness (left column, top sub-panel), percentage of spurious sources (left column, bottom sub-panel), and flux density comparison (right column) between the input sources and the recovered ones by flat (blue series) and spectral (red series) MultiPoSeIDoNs, the MTXFs with a detection limit of 60 mJy (cyan series), and with a 4σ detection limit (orange series). The frequencies are 143 (top panels), 217 (middle panels), and 353 GHz (bottom panels) for a 30° Galactic cut. The dotted grey vertical line is the 90% completeness flux density limit (177, 152, and 304 mJy for 143, 217, and 353 GHz, respectively) for the second Planck catalogue of compact sources (Planck Collaboration XXVI 2016).

In the text
thumbnail Fig. 4.

Validation results of completeness (left column, top sub-panel), percentage of spurious sources (left column, bottom sub-panel), and flux density comparison (right column) between the input sources and the recovered ones by Flat (blue series) and Spectral (red series) MultiPoSeIDoNs and PoSeIDoN (green series). The frequency is 217 GHz for a 30° Galactic cut. The dotted grey vertical line is the 90% completeness flux density limit (152 mJy) for the second Planck catalogue of compact sources (Planck Collaboration XXVI 2016).

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.