Issue |
A&A
Volume 673, May 2023
|
|
---|---|---|
Article Number | A103 | |
Number of page(s) | 16 | |
Section | Numerical methods and codes | |
DOI | https://doi.org/10.1051/0004-6361/202245750 | |
Published online | 16 May 2023 |
The miniJPAS survey quasar selection
III. Classification with artificial neural networks and hybridisation
1
Instituto de Astrofísica de Andaluciá (CSIC),
PO Box 3004,
18080
Granada,
Spain
e-mail: gimarso@iaa.es
2
Departamento de Astronomia, Instituto de Física, Universidade Federal do Rio Grande do Sul (UFRGS),
Av. Bento Gonçalves,
9500,
Porto Alegre, RS,
Brazil
3
Departamento de Física Matemática, Instituto de Física, Universidade de São Paulo,
Rua do Matão, 1371,
CEP 05508-090,
São Paulo,
Brazil
4
Sorbonne Université, Université Paris-Diderot, CNRS/IN2P3, Laboratoire de Physique Nucléaire et de Hautes Energies, LPNHE,
4 Place Jussieu,
75252
Paris,
France
5
Institut de Física d’Altes Energies (IFAE), The Barcelona Institute of Science and Technology,
08193
Bellaterra (Barcelona),
Spain
6
Aix-Marseille Univ., CNRS, CNES, LAM,
Marseille,
France
7
Donostia International Physics Center,
Paseo Manuel de Lardizabal 4,
20018,
Donostia-San Sebastian (Gipuzkoa),
Spain
8
Centro de Estudios de Física del Cosmos de Aragón (CEFCA),
Plaza San Juan, 1,
44001
Teruel,
Spain
9
Centro de Estudios de Física del Cosmos de Aragón (CEFCA), Unidad Asociada al CSIC,
Plaza San Juan, 1,
44001
Teruel,
Spain
10
Donostia-San Sebastian, Spain Ikerbasque, Basque Foundation for Science,
48013
Bilbao,
Spain
11
Department of Astronomy, University of Illinois at Urbana-Champaign,
Urbana, IL
61801,
USA
12
INAF, Osservatorio Astronomico di Trieste,
via Tiepolo 11,
34131
Trieste,
Italy
13
IFPU, Institute for Fundamental Physics of the Universe,
via Beirut 2,
34151
Trieste,
Italy
14
Observatório Nacional, Rua General José Cristino,
77, São Cristóvão,
20921-400
Rio de Janeiro,
Brazil
15
Department of Astronomy, University of Michigan,
311 West Hall, 1085 South University Ave.,
Ann Arbor,
USA
16
Department of Physics and Astronomy, University of Alabama,
Box 870324,
Tuscaloosa, AL,
USA
17
Universidade de São Paulo, Instituto de Astronomia, Geofísica e Ciências Atmosféricas,
R. do Matão 1226,
05508-090
São Paulo,
Brazil
18
Instruments4,
4121 Pembury Place,
La Cañada-Flintridge, CA
91011,
USA
Received:
21
December
2022
Accepted:
14
March
2023
This paper is part of large effort within the J-PAS collaboration that aims to classify point-like sources in miniJPAS, which were observed in 60 optical bands over ~1 deg2 in the AEGIS field. We developed two algorithms based on artificial neural networks (ANN) to classify objects into four categories: stars, galaxies, quasars at low redshift (z < 2.1), and quasars at high redshift (z ≥ 2.1). As inputs, we used miniJPAS fluxes for one of the classifiers (ANN1) and colours for the other (ANN2). The ANNs were trained and tested using mock data in the first place. We studied the effect of augmenting the training set by creating hybrid objects, which combines fluxes from stars, galaxies, and quasars. Nevertheless, the augmentation processing did not improve the score of the ANN. We also evaluated the performance of the classifiers in a small subset of the SDSS DR12Q superset observed by miniJPAS. In the mock test set, the f1-score for quasars at high redshift with the ANN1 (ANN2) are 0.99 (0.99), 0.93 (0.92), and 0.63 (0.57) for 17 < r ≤ 20, 20 < r ≤ 22.5, and 22.5 < r ≤ 23.6, respectively, where r is the J-PAS rSDSS band. In the case of low-redshift quasars, galaxies, and stars, we reached 0.97 (0.97), 0.82 (0.79), and 0.61 (0.58); 0.94 (0.94), 0.90 (0.89), and 0.81 (0.80); and 1.0 (1.0), 0.96 (0.94), and 0.70 (0.52) in the same r bins. In the SDSS DR12Q superset miniJPAS sample, the weighted f1-score reaches 0.87 (0.88) for objects that are mostly within 20 < r ≤ 22.5. We find that the most common confusion occurs between quasars at low redshift and galaxies in mocks and miniJPAS data. We discuss the origin of this confusion, and we show examples in which these objects present features that are shared by both classes. Finally, we estimate the number of point-like sources that are quasars, galaxies, and stars in miniJPAS.
Key words: methods: data analysis / surveys / galaxies: Seyfert / quasars: emission lines / cosmology: observations
© The Authors 2023
Open Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This article is published in open access under the Subscribe to Open model. Subscribe to A&A to support open access publication.
1 Introduction
The new era in modern astronomy is closely connected to the era of big data. Astronomical observations produce increasingly larger amounts of data. The new generation of surveys, such as the Dark Energy Spectroscopic Instrument (DESI; Levi et al. 2013), the Large Synoptic Survey Telescope (LSST; Ivezić et al. 2019) or the Square Kilometer Array (SKA; Dewdney et al. 2009), will observe of the order of millions or even billions of objects. In particular, the Javalambre Physics of the Accelerating Universe Astrophysical Survey (J-PAS; Benitez et al. 2014) will observe thousands of deg2 in the northen sky in the upcoming years in 54 narrow-band filters, detecting more than 40 million objects. Consequently, it is necessary to automatise all the tasks as much as possible so as to process the astronomical information faster and more efficiently. The identification and classification of astronomical objects is certainly the first step prior to any further scientific analysis.
Traditionally, photometric surveys identified galaxies and stars based on their morphological structure and colour properties (see e.g. Baldry et al. 2010; Henrion et al. 2011; Saglia et al. 2012; López-Sanjuan et al. 2019). Typically, galaxies are extended objects, while point-like sources are mainly either stars or quasi-stellar objects (QSOs). Nevertheless, the lack of spatial resolution for the most distant and faint galaxies causes them look very similar to point-like sources. Furthermore, the colour space in multi-band photometric surveys becomes increasingly complex as the number of filters increases, which requires more sophisticated algorithms to fully exploit all the information encoded in the surveys.
In past years, machine-learning (ML) algorithms have been used in many applications within the astronomical field, from the estimation of photometric redshifts (Cavuoti et al. 2017; Pasquet et al. 2019; Ramachandra et al. 2022), the identification of low-metallicity stars (Whitten et al. 2019), the determination of the star formation rate (Delli Veneri et al. 2019; Bonjean et al. 2019), the classification of morphological types in galaxies (Domínguez Sánchez et al. 2018), and the identification of causality in galaxy evolution (Bluck et al. 2022) to the measurement of the equivalent widths of emission lines in photometric data (Martínez-Solaeche et al. 2021). The problem of source identification in photometric surveys has also been addressed by ML either to distinguish between point-like and extended sources (Vasconcellos et al. 2011; Kim et al. 2015; Kim & Brunner 2017; Burke et al. 2019; Baqui et al. 2021) or even between galaxies, stars, and QSOs (Krakowski et al. 2016; Bai et al. 2019; Logan & Fotopoulou 2020; Xiao-Qing & Jin-Meng 2021; He et al. 2021).
The goal of the present paper is to classify the objects detected with the miniJPAS survey (Bonoli et al. 2021) into stars, galaxies, QSOs at high redshift (z ≥ 2.1), and QSOs at low redshift (z ≤ 2.1) by using artificial neural networks (ANN). The threshold at z = 2.1 corresponds to the limit at which QSOs show the Lyman-α emission line within the J spectra. The miniJPAS survey is part of the J-PAS project1, which detected more than 60 000 objects within the All-wavelength Extended Groth Strip International survey (AEGIS; Davis et al. 2007) using 56 narrow-band J-PAS filters (~145 Å) and the four ugri broad-band filters. The separation of 100 Å among filters makes the J-PAS filter system equivalent to obtaining a low-resolution spectrum with R ~ 60 (J spectrum hereafter). These unique characteristics enable observation and analysis of galaxies and QSOs in continuous redshift ranges, 0 ≲ z ≲ 1 and 0.5 ≲ z ≲ 4, respectively (Bonoli et al. 2021). Different studies have proved the capability of J-PAS to address several topics within the astrophysical field, for instance, the evolution of the stellar population properties of galaxies up to z ~ 1 (González Delgado et al. 2021), the properties of the nebular emission lines of galaxies down to z ≤ 0.35 (Martínez-Solaeche et al. 2022), the measurement of black hole virial masses for the QSO population (Chaves-Montero et al. 2022), or the study of galaxy properties within galaxy clusters (Rodríguez-Martín et al. 2022) and groups (González Delgado et al. 2022). Unfortunately, the data available for spectroscopically confirmed sources in the miniJPAS area are not sufficient to train and test ML algorithms for our purpose. Therefore, we employed mock data developed by Queiroz et al. (2022), and used the sources identified spectroscopically within the AEGIS field by the Sloan Digital Sky Survey (SDSS; York et al. 2000) in the DR12Q superset catalogue (Pâris et al. 2017) as truth table.
Modern deep ANN generally perform better than traditional methods. They remain poorly calibrated most of the time, however (Goodfellow et al. 2014; Guo et al. 2017). The probabilities associated with the predicted label classes may be overly confident because they do not correspond to true likelihoods. Consequently, objects are classified as part of one class or another with high probability regardless of the prediction accuracy. Realistic probability distributions are particularly important for spectroscopy follow-up programs that typically prioritise the observation of high-probability objects of some particular class. Furthermore, some objects are indeed dual in nature. Although we consider as QSOs only galaxies with an extremely luminous active galactic nucleus (AGN), there are objects in which a significant fraction of the detected light comes from the stars in the host galaxy. In this scenario, ML algorithms should ideally provide a high probability in both classes.
In a recent paper, Zhang et al. (2017) proposed an original idea called mix up to enlarge the dataset and improve the generalisation of the trained model, thus increasing the robustness to adversarial examples. Later on, Thulasidasan et al. (2019) showed that mix up or hybridisation, as we prefer to call it, also improves the calibration and predictive uncertainty of deep neural networks. In this work, we enlarge our training set by mixing features from stars, galaxies, and QSOs, and we study the effect of hybridisation on the performance and calibration of the models. The ML classifiers used in this work will be combined with other ML algorithms. In Rodrigues et al. (2023), we trained convolutional neural networks (CNNs) by proposing different approaches to incorporate photometry uncertainties as inputs in the training phase. In this work, we focus our attention on the galaxy-QSO degeneracy and the ability of ANN to estimate realistic probability density distributions (PDF). In Rodrigues et al. (2023), we study other relevant aspects, such as the stellar types that are more frequently confused with the QSO population, the J-PAS feature importance, or the stability of the CNNs predictions with respect to minor changes in the training set. In Pérez-Ràfols et al. (in prep. a), we additionally adapt the code SQUEZE (Pérez-Ràfols et al. 2020) to work with J-PAS data. This code is based on optical emission line identification to separate between QSOs and non-QSOs and also estimates the redshift of QSOs. Ultimately, all these codes will be merged in a combined algorithm (Pérez-Ràfols et al., in prep. b) so as to classify the miniJPAS sources more efficiently and to provide a high-redshift QSO target list for a spectroscopic follow-up with the WEAVE multi-object spectrograph survey (Dalton et al. 2014; Jin et al. 2023), which is planning to carry out a Lyman-α forest and metal line absorption survey (Pieri et al. 2016).
This paper is organised as follows. In Sect. 2, we present miniJPAS data, and we briefly summarise the processes employed in the construction of the mock catalogue. In Sect. 3, we describe the main characteristics of the classifiers in detail and relate how data augmentation was employed through hybridisation. We indicate the performance metrics used for testing purposes in the paper in Sect. 3.5, and we show the main results obtained in the paper in Sects. 4 and 5. Finally, we summarise and conclude in Sect. 6. Throughout the paper, all magnitudes are presented in the AB system (Oke & Gunn 1983).
2 MiniJPAS survey and mocks
The miniJPAS survey includes data from four pointings scanning ~1 deg2 along the AEGIS field. The photometric system includes 56 bands, namely 54 narrow-band filters in the optical range plus two medium bands, one in the near-UV (uJAVA band), and the other in the NIR (J1007 band). With a separation of ~100 Å, each narrow-band filter has a full width at half maximum (FWHM) of ~145 Å, whereas the FWHM of the uJAVA band is 495 Å, and J1007 is a high-pass filter. Additionally, four broad bands u,g,r, and i were used to complement the observations. These were carried out with the 2.55 m telescope at the Observatorio Astrofísico de Javalambre, a facility developed and operated by CEFCA, in Teruel (Spain). The data were acquired using the pathfinder instrument, a single CCD direct imager (9.2k × 9.2k, 10µm pixel) located at the centre of the 2.55 m telescope FoV with a pixel scale of 0.23 arcsecpix−1, vignetted on its periphery, providing an effective FoV of 0.27 deg2. The r band was chosen as the detection band, and the reference image is in the dual-mode catalogue. This image was used to define the position and sizes of the apertures from which the rest of the photometry was extracted. The miniJPAS survey is 99% complete up to r ≤ 23.6 mag for point-like sources, and it detected more than 60 000 objects2 (Bonoli et al. 2021). In this paper, we only analyse objects with FLAGS= 0 and MASK_FLAGS= 0 (46441 in total), that is, they are free from detection issues such as contamination from bright stars and light reflections in the telescope or in its optical elements. We refer to Bonoli et al. (2021) for details on the flagging scheme. Removal of flagged sources from the catalogue decreased our sample size, but did not introduce any bias because the fraction of sources that are flagged is independent of their magnitude (Hernán-Caballero et al. 2021).
The algorithms presented in this work were trained and tested on mock data (Queiroz et al. 2022). The J spectra of galaxies, stars, and QSOs were simulated by convolving SDSS spectra included in the SDSS DR12Q superset catalogue (Pâris et al. 2017) with the transmission profiles of J-PAS photometric system (synthetic fluxes). The SDSS DR12Q superset contains all objects targeted as QSOs from the final data release of the Baryon Oscillation Spectroscopic Survey (BOSS; Dawson et al. 2013). Therefore, it also contains galaxies and stars whose broadband colours are compatible with those from QSOs. Since the SDSS sample is complete only up to r ~ 20.5, fainter objects were generated by adding random fluctuations (noise) to the synthetic fluxes. The mock catalogue includes several noise models that mimic the observed signal-to-noise ratio (S/N) in miniJPAS for the APER_3 magnitudes. The APER_3 magnitude contains the light integrated within an aperture of three arcsec. Thus, in order to account for the missing light emitted by each source, we implemented an aperture correction that was based on the APER_6 magnitude, as detailed in Queiroz et al. (2022). The main results presented in this paper are not affected by the noise model choice. We used model 1, which assumes that the noise distribution of miniJPAS point-sources in each filter is well described by a single Gaussian distribution (Queiroz et al. 2022). We considered SDSS spectra to be noise free. Although the S/N is lower at short and long wavelengths, the error in the mocks is dominated by miniJPAS observations, which reach fainter objects. As we did not simulate miniJPAS images, the photometry in the mocks is not affected by any image-processing effect. This choice makes our mocks different from observations to a certain extent. For example, miniJPAS observations were not carried out with 56 filters simultaneously. Instead, the observations were taken in trays of 14 CCDs, thus different observational conditions were presented in the SED of individual objects. Taking these effects into consideration would have made the mock-building process very complex and tedious. Additionally, there is no guarantee that these modifications would have improved the performance of the algorithms. In the future, as soon as J-PAS begins to observe the sky, we will have J-PAS data for objects that were previously observed by other spectroscopic surveys such as WEAVE. We will then be in a position to re-train our models with observed data and reduce the existing gap between simulated and real observations.
Pérez-Ràfols et al. (in prep. b) will compare the performance of each algorithm in the mock test sample with different noise models, and we will classify sources in miniJPAS for each one of them. Galaxies follow the magnitude-redshift distribution of SDSS and DEEP3 (Cooper et al. 2011, 2012) found in miniJPAS. QSOs follow the luminosity function of Palanque-Delabrouille et al. (2016), and stars are distributed according to the Besançon model of stellar population synthesis of the Galaxy (Robin et al. 2003) and the SDSS miniJPAS spectroscopic sample.
The number of stars, galaxies, and QSOs in training set are balanced to prevent biases in the classifiers towards over-represented classes. The same applies for the test set and the validation set, which were used to fine-tune the hyper-parameter of the classifiers. Additionally, another test sample, the 1-deg2 test set, includes the expected number of point-like sources within the miniJPAS area. This sample was generated to provide a more direct comparison with miniJPAS observations. Finally, we used a sub-sample of miniJPAS observations as a true table. This sample is called the SDSS test sample and includes only the objects observed by the SDSS DR12Q superset. Unfortunately, the SDSS test sample does not contain a sufficient number of objects, and it is biased toward bright objects. Therefore, it can only give us a small impression of the actual performance of the algorithms in real data. In Table 1 we summarise the number of objects contained in all samples. Further details of how these synthetic data were created can be found in Queiroz et al. (2022).
Number of objects in each data set.
3 Star, galaxy, and QSO classifier
In this section, we describe in detail how the ML classifiers were developed. Although our main focus is on ANN, we also developed an RF classifier in order to compare the performance of the two algorithms. They were designed to distinguish between four different classes: stars, galaxies, QSOs at high redshift (z ≥ 2.1), and QSOs at low redshift (z < 2.1), referred to as QSO-h and QSO-l, respectively.
3.1 Artificial neural networks
The ANN were coded with the Tensorflow (Abadi et al. 2016) and Keras libraries (Chollet et al. 2015) in Python. The ANN has eight hidden layers with 200 neurons each. As a regularisation technique, we used weight constraints and imposed a maximum value of two in each neuron (kernel constraint). We also removed 15% of the neurons in each layer. We used the rectified linear unit (ReLU) as our activation function (Nair & Hinton 2010). Weights were initialised with the He initialisation strategy (Géron 2019). The loss function we employed is the cross entropy.
We trained two models that use two different sets of inputs, called ANN1 and ANN2. The inputs of ANN1 (59 in total) were relative fluxes, that is, the flux in each filter was divided by the flux in the r band. Since the miniJPAS dual-mode catalogue used the r band for detection, this normalisation is well defined for all the objects. The inputs of ANN2 were the colours measured with respect to the mAB in the r band, plus the normalised magnitude in this band (60 inputs in total), (1)
where mAB and fλ stand for the magnitude and the flux in the jth filter, respectively, and min() and max() are the minimum and maximum values of the magnitude in the r band within our training set. Both sets of inputs capture the shape of the spectrum, but the ANN2 also includes information about the observed luminosity of each source, which anchors the SED to a particular magnitude.
Objects in the dual-mode catalogue might be undetected in some bands (non-detection). This happens when the S/N values for these bands are very low and the measured fluxes are null or negative after the sky background subtraction. In the mock catalogues, a non-detection (ND) follows the pattern observed in miniJPAS. For specific details of how ND are modelled, we refer to Queiroz et al. (2022). We set the inputs of the ANN2 to zero when fluxes are negative because the colours are otherwise undefined. This approach in sub-optimal in the sense that colours might be zero if fluxes are identical, thus it is difficult for the ANN to identify NDs. Another option is to set NDs to an arbitrary number. Unfortunately, no improvement has been observed with this approach. In the case of ANN1, we allowed the inputs to be below zero because values are positive by definition. In principle, the ANN1 should be able to model the sky background better because it can identify the regime of low emission in which fluxes are close to zero or even negative.
3.2 Random forest
Random forest (RF) is an ensemble machine-learning algorithm that uses multiple decision trees to predict an outcome. Multiple trees (a forest) are built using a random subset of features and observations from the training data (bootstrapped sample). For each bootstrapped sample, a decision tree is grown to the maximum depth or until a stopping criterion is reached. The effectiveness of each split in a decision tree can be quantified by the information gain, which is a measure of the expected reduction in impurity as defined by either entropy or the gini impurity function. The intrinsic randomness of the RF helps reduce the risk of overfitting and creates a more robust model than single decision trees. After building several trees, the algorithm takes the average or majority vote of the predictions from each tree to produce a final prediction for each data point.
We built an RF classifier that uses the same features as ANN1 and ANN2 in order to compare them. The algorithm was implemented with the scikit-learn python package with the default setting3.
3.3 Data augmentation via hybridisation
Data augmentation has been proven to be an excellent tool to increase the size of the training sample, and consequently, the performance of ML algorithms when only limited training samples are available (Shorten & Khoshgoftaar 2019). Rotation, translation, or scaling are among the most popular techniques for image classification (Yang et al. 2022). In the case of nonimage features such as the J spectra, the most common manner to perform data augmentation is via Gaussian noise. However, the benefit of this technique in our training sample would be limited because it was already used to generate objects at different magnitudes bins in the construction of the mock catalogue. Thus, we adapted the mix up technique proposed in Zhang et al. (2017) to our classifiers, which aim to distinguish between four classes. This technique allowed us to enlarge the training set by mixing features from different classes, generating a new training set composed only of hybrid objects. The new set of hybrid objects () and their respective fluxes were generated as a linear combination of individual objects in the original training set, (2) (3)
where y1 (f1(λ)), y2 (f2(λ)), y3 (f3(λ)), and, y4 (f4(λ)) are the vectors (fluxes) of each one of the classes (stars, galaxies, QSO-l, and QSO-h, respectively), and αj is the mixing coefficient, which varies between zero and one according to an exponential distribution function that depends on β the scale parameters that control the level of mixing, (4)
where x is a positive random variable between 0 and 1. As ß increases, αj is more likely to become smaller, thus more mixing (1 − αj) is performed between classes. Finally, in Eq. (3), δij is the Kronecker delta, and cij are the luminosity coefficients, given by (5)
where Ni is the number of objects belonging to class i within each magnitude bin in the original training set, and Ntot = N1 + N2 + N3 + N4. The luminosity coefficients modulate how much of the mixing (1 − αj) comes from each class as a function of their relative number in different magnitude bins. For instance, if we generate an hybrid galaxy () at magnitude , Eq. (3) becomes (6)
For low values of β, α2 is near one with a high probability. Therefore, the new hybrid galaxy is still a galaxy, but it has some of level of contamination from the other classes. The probability of not being a galaxy (1 − α2) is distributed among the other classes by taking their relative amounts at into account. Since stars are less frequent at this magnitude, c21 is close to zero, and the new hybrid galaxy is mixed mainly with QSO-l and QSO-h. In order to compute the cij coefficients, we split the training set into rSDSS magnitude bins that contained roughly 20 000 objects. In this way, only objects with a similar brightness were mixed together. After hybridisation, the proportion of each class in the training set remained the same. That is to say, even though features of different objects were mixed together, each object conserves most of the original features, and hence they still have a high probability to belong to the original class. We enlarged the training set five times with ß = 0.1 (we discuss this choice in Sect. 4.1). We caution that hybridisation does not mix objects following a physical recipe, but is rather a mathematical transformation of the data. Since hybridisation was implemented at the level of fluxes, Eq. (1) still needs to be applied to the resulting hybrid fluxes in order to standardised the input data for ANN1 and RF1 , and RF2 and ANN2.
3.4 Training strategy
The intrinsic randomness of the training procedure usually leads to solutions that are not optimal. Weights and biases are drawn from a distribution function that generates the initial state. Therefore, each time that the training is performed, the algorithm converges to a different local minimum of the loss function. Furthermore, the training set itself is augmented in a random manner via hybridisation. The hybrid space is filled in a slightly different way in each realisation. In the limit case where the hybrid set is much larger than the original, this effect would be negligible. However, a huge training set is less practical to handle and more difficult to train than a smaller set. For all these reasons, we followed the committee approach (Bishop 1995), that is, we trained several ANNs and computed the median to provide a final classification. Then, we re-normalised the output probabilities to ensure that the sum was one. In order to determine the optimal number of ANN or committee members needed, we started with two, and we added an additional member at each step until the results did not improve for the validation sample in terms of the f1 score (see Sect. 3.5 for a definition of this metric). We determined that eight members are enough to reach convergence.
3.5 Performance metrics
In this section, we provide a detailed discussion of the metrics used to evaluate the performance and robustness of the classifiers. Each metric offers unique insights into the ability of the ML algorithms to classify sources. By examining these metrics, we can better understand the strengths and limitations of the classifiers.
3.5.1 Confusion matrix
The confusion matrix is especially useful in the context of a multi-label classification problem. The actual classification of each object is shown in the columns of the matrix, while the predicted objects lie in the rows. Therefore, in the best (ideal) case scenario, the matrix would be purely diagonal, and every prediction would coincide with the actual classification. Non-diagonal terms indicate which classes are confused and provide valuable information for improving the training set and fine-tuning the hyperparameters of the model.
3.5.2 f1-score
Unlike the confusion matrix, the f1-score yields one single scalar for each one of the classes. It finds a compromise between purity (precision) and completeness (recall), (7) (8)
where TP, FP, and, FN are the true-positive rate, the false-positive rate, and the false-negative rate, respectively. FP appears as non-diagonal terms in the columns of the confusion matrix, while TN lies on the rows. In the case of an unbalanced test set in which one or more classes are underrepresented, the average performance of the model can be estimated using the weighted f1 score, (9)
where ni is the number of objects belonging to the ith class in the test set, and nobj is the total number of objects.
3.5.3 Expected calibration error
A well-calibrated probabilistic classifier is able to predict probabilities that coincide on average with the fraction of objects that truly belong to a certain class (accuracy). We assume that we took 100 objects with a probability of 10% of being a star. When the classifier is well calibrated, about 10 of them should be stars. If there are more, our classifier is under-confident. If there are less, then the classifier is over-confident. Probability calibration curves are normally employed to display this relation, where we bin the probability estimates and plot the accuracy versus the mean probability in each bin. Let Bmj be the set of objects whose predicted probabilities of being class j fall into bin m. The accuracy and confidence of Bmj are defined as (10) (11)
where Pij is the probability of being class j for the ith object, and Nmj is the number of objects of class j within bin m. The expected calibration error (ECE) is then defined as (12)
where Nc is the number of classes, and Nb is the number of bins. The lower the value of the ECE, the better the calibration of the model. However, the output of the ANN only represents true probabilities under the assumption that our mock sample and the miniJPAS observations do not differ essentially. In order to have a better estimate of the ECE of the model, we would need to compute this metric in a sufficiently large true table, which is not yet possible. In the future, when more data will be gathered, we will be able to employ this metric on observations and evaluate the ECE of the ANN properly. In the remainder of the paper, we refer to the outputs of the ANN as probabilities, recalling that they are simply a proxy of the true probabilities.
3.5.4 Entropy
The entropy is a measurement of disorder. In the context of ML, the entropy of a classifier prediction can tell us how uncertain the classifier is. The entropy of the ith object can be written as follows: (13)
where Pij is the probability that the ith object is class j, Nc is the number of classes, and ϵ is an arbitrarily small number (10−14) to avoid the divergence of the logarithm when the probability for a given class is zero. The entropy is maximum (log2 Nc) when each class has a probability of 1/Nc and zero when the probability of belonging to a particular class is one.
4 Results
In this section, we test the performance of the algorithms on the mock test sample (Sect. 4.1). We discuss in detail the effect of augmenting the data through hybridisation and we compare the differences between classifiers. Finally, we evaluate the classification obtained with SDSS objects observed by miniJPAS in the AEGIS field.
Fig. 1 score for different magnitude bins as defined in Eq. (14), and the score for the full sample (ALL BIN). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix are trained with fluxes (see Sect. 3.1). |
4.1 Test sets
The performance of a classifier changes as a function of the magnitude of the objects. Fainter objects are more difficult to classify because their S/N is lower. In order to quantify the potential bias of the classifiers at different magnitudes, we therefore split the validation and the test samples into three different bins according to the r-band magnitude, (14)
The number of objects in the test sample for BIN 0, 1, and 2 are 4987, 13 357, and 9488, respectively. In Fig. 1, we show the score for each of the magnitude bins defined above, including the average performance for the full sample (ALL BIN). We compare the score of the ANN trained with the hybrid set (ANN1 mix and ANN2 mix) and with the original training set (ANN1 and ANN2). In Fig. 2, the ƒ1 score is also shown for each of the classes. Overall, both classifiers (ANN1 and ANN2) are very similar, with small differences in each magnitude bin for each class. The ANN1 classifier is slightly better probably not only because of the representation of the data (relative fluxes), but also because it captures the sky background better.
As expected, the accuracy of the classifiers decreases for fainter objects. The performance obtained with the hybrid set is very similar to the original training set, suggesting that the latter already contains all the variance needed, and more examples do not necessarily imply a better performance (but see the next section for a more detailed discussion).
The confusion matrices as a function of the magnitude bin for the ANN1 model are shown in Fig. 3. The matrices from the remaining models are provided in Appendix A. QSO-1 and galaxies are more difficult to distinguish, especially at the faint end, where the data are noisier. These objects do not belong to independent classes. Sometimes, the host galaxy of an AGN might contribute strongly to the SED. In Seyfert galaxies, the observed spectrum is usually a combination of the light coming from the AGN and the stellar populations within the galaxy. Therefore, we expect confusions between QSO-1 and galaxies more often than between any of the other classes. Finally, in the faintest magnitude bin, 31.4% of QSO-h are classified as QSO-1, and 18.9% of the stars are confused with QSO-1.
In Fig. 4, we show examples of the most common missclassifications. The first row is composed of QSO-1 that were classified as galaxies. In the second row, we show galaxies that were identified as QSO-1, the third row corresponds to QSO-h confused with QSO-1, and the last row shows stars classified as QSO-1. Even though we were unable to correctly predict the class of these objects, the second most likely class usually coincides with the actual class. Furthermore, it is important to emphasise that the objects shown in Fig. 4 would be very difficult to identify via visual inspection even for a human expert without seeing the spectrum. ML algorithms are indeed pushing the limits beyond the human capability.
It is expected that the ANN predictions for low S/N objects are more uncertain than those with high S/N. In Fig. 5, we show the median entropy as a function of the median S/N in bins of 1000 objects. While the predictions are highly certain in the high S/N limit with the ANN1 and ANN2 classifiers, the entropy obtained by the ANNs trained in the hybrid sets remains almost constant from an S/N of 25 to 10 with a value of ~0.4, and then it increases slightly. The lowest entropy obtained with the ANN1 mix and ANN2 mix classifiers is governed by the mixing coefficient (β) used to generate the hybrid set in Eq. (3), and it coincides with the median entropy of the hybrid classes in the training set. In Sect. 3.3, we chose β to be 0.1. This choice was arbitrary, but it does not have a major impact on the performance of the algorithms. Mainly the entropy (and probabilities) of the classifier solution is affected. On the one hand, as β increases, the entropy of the hybrid training set increases up to a maximum value at which all objects reach equal probabilities. For these values, we expect the performance to be degraded because the class information completely vanished. On the other hand, for very low values of ß, we recover the original training set, thus no hybridisation is performed in practice. In essence, no optimal value for ß exists that might improve the classifier performance and confidence simultaneously.
In Fig. 6, we show the fraction of positive detection for each one of the classes as a function of the mean probability obtained with the ANNs in the mock test sample (r ≤ 23.6). In the top (bottom) left panel, we show the results of the ANN1 (ANN2) predictions trained with the original training set, and in the right panel, we show the predictions obtained with the hybrid set. The ECEs for galaxies, QSO-h, QSO-l, and stars are shown in the top left panel. Training with hybrid classes has a negative impact on the calibration. Once again, the ECE is a function of the mixing coefficient: as α increases, the ECE increases. Overall, the ANN1 is slightly better calibrated than the ANN2, but ANN2 mix is better than ANN1 mix.
It is worth considering whether hybridisation improves the performance of the ANNs when the training set is smaller. The original training set in the mock catalogue is composed of 300 000 objects, and the hybrid set is five times larger. We assumed a training set that was ten times smaller than the original set (reduced set). After applying hybridisation, we generated two new training sets that were five and ten times larger than the reduced set, respectively, known as the reduced hybrid set ×5 and the reduced hybrid set × 10. We then compared the performance of ANN1 in the mock test sample. In Fig. 7, we show the difference between the score in each one of the mentioned training sets and the score obtained in the original training set as a function of the median S/N. We do not observe a significant improvement that might justify the use of hybridisation, at least in the form we implemented in Eq. (3) for this particular data set.
Finally, we compared the performance of ANNs with the results of RF classifier. In Fig. 8, we show the weighted score as a function of the median S/N in the observed filters for all classifiers. Each bin contains roughly 1000 objects. We do not show the results of ANN1 mix and ANN2 mix because they are very similar with respect to their non-hybrid counterparts. The ANNs perform better and score 10% higher on average than the RF classifiers. It is remarkable that even with a median S/N of 5, the score reaches 0.9. We do not observe a significant difference between RF1 and RF2 or ANN1 and ANN2, which suggests that both representation of the data encode essentially the same information. Hybridisation seems to improve the performance of the RF slightly. Nevertheless, the ECE increases from 3.89 (3.82) up to 4.23 (4.19) for RF and RF2, respectively. The ECE obtained with the original training set is much larger than what we found with ANN. Thus, the RF classifier tends to be under-confident in its predictions.
Fig. 2 f1 score for each of the classes galaxies, QSO-h, QSO-1, and stars as a function of the magnitude bins defined in Eq. (14), and score for the full sample (ALL BIN). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix do with fluxes (see Sect. 3.1). |
Fig. 3 Confusion matrices obtained with the ANN1 in the test sample. |
4.2 SDSS versus miniJPAS
In this section, we test the ANN classifiers on the SDSS test sample. Figure 9 shows the f1 score for each class and the score. The performance of the algorithms that were trained with the hybrid set (ANN1 mix and ANN2 mix) are compared with the original training set (ANN1 and ANN2). Due to the limited number of objects, we did not separate these samples into magnitude bins. Most of the objects (75%) belong to BIN 1, and only three are at the faint end (BIN 2). Therefore, the f1 score is mostly representative of BIN 1. The results of the SSDS test sample are compatible with those obtained for the mock data (~0.9), suggesting that the simulations reproduce the miniJPAS observations fairly well at least for magnitudes brighter than 22.5. Unfortunately, we do not have enough labelled objects fainter than 22.5 within the miniJPAS field. We therefore needed to rely on the mock results for an expectation of the performance. As soon as WEAVE starts to observe the QSO target list provided by the J-PAS, we will be able to fully assess the performance of the algorithms for the full range of magnitudes.
In Fig. 10, we show the confusion matrix obtained with ANN1 for the SDSS test sample. The matrix for BIN 0 is purely diagonal and includes only one confusion between a low-redshift quasar and a galaxy (object ID: 2406-15603), which is shown in Fig. 11 and is discussed below. Thus, the objects confused by the algorithms are all from BIN 1. The confusion matrices for the remaining models are listed in Appendix B. The sample of QSO predicted by the ANN, and especially the sub-sample of QSO-h, contains very few false positives (QSO columns in the confusion matrix), meaning that the algorithms favour a pure rather than a complete sample. However, the sample of galaxies is more complete because ~ 19 % of them are classified as QSO-l, but only a few SDSS galaxies are missing. Finally, stars are identified with a very high accuracy, with only five false positives and eight true negatives. We recall that these results are partially biased due to the small number of objects in it and the selection criteria that were used by the SDSS team to select the sample of QSO targets. Hence, it can only give us a glimpse of the actual performance of the ANN in real data.
In Fig. 11, we show some examples of objects that were observed simultaneously by SDSS and miniJPAS below redshift 1 that might present a spectrum composed of mixed features. In other words, the light coming from these objects has received contributions from both the AGNs and the stellar populations within the galaxies. All the objects except 2470-3341 are classified in SDSS as QSO-l. However, SExtractor identified them as extended sources with a class-star value below 0.35 (CL in Fig. 11). Following the SDSS classification criterion, only 2470-3341 and 2241-1234 are correctly classified by the ANN1. Nevertheless, 2406-15603 is rather a Seyfert 1 galaxy with broad emission lines such as Hα and Hß. It also has a reddish spectrum, and the extended structure of the galaxy is very clearly visible in the image. Furthermore, while the SDSS spectrum detects the broad emission line of Hα, the miniJPAS observation does not capture this emission, which is probably the most relevant feature for classifying this object as a QSO. Sources 2241-18615, 24062560, and 2406-7300 are classified as galaxies, but the second preferred class is QSO-l. These objects are indeed not very different from 2470-3341, which is a Seyfert 2 galaxy according to the SDSS pipeline. Once again, the J spectra miss two relevant features in 2241-18615 and 2406-2560, the Hα and Hß emission lines, respectively. Finally, 2241-1234 is correctly classified. Although it is an extended source according to SExtractor, the emission of the AGN dominates the spectrum. The high S/N obtained in this object makes the classification more certain.
Fig. 4 Examples of the most typical misclassification of objects in the mock test sample. The first row shows QSO-l classified as galaxies, the second row shows galaxies classified as QSO-l, the third row shows QSO-h classified as QSO-l, and the fourth row shows stars classified as QSO-l. From left to right, the objects are fainter. In each panel, we indicate the AB magnitude in the r band, the redshift (top left), and the probabilities yielded by the ANN1 classifier for each one of the classes (top right). The SDSS spectra from which the test sample is generated are shown in the background in grey. |
Fig. 5 Median entropy as a function of the median S/N in bins of 1000 objects. |
5 Star, galaxy, and QSO classification in miniJPAS
We now focus our attention on the classifier predictions on miniJPAS data. In Fig. 12, we show the distribution of the confidence (probability) levels yielded by the ANN1 and ANN2 classifiers for each class and each magnitude bin. We only predict the class for the objects that are considered point-like sources according to SExtractor, which used the detection band (rSDSS). Both ANN classifiers predict roughly the same number of objects, but they exhibit differences in the faintest magnitude bins, which are useful to later build a combined algorithm that uses information from several classifiers (Pérez-Ràfols et al., in prep. b).
The total number of QSOs predicted by ANNs for miniJPAS is compatible with previous estimates. Abramo et al. (2012) expected ~240 QSOs per square degree with the J-PAS photometric system for a limiting magnitude of g = 23, assuming that QSOs follow the luminosity function found by Croom et al. (2009) and the QSO selection is perfect. With the ANN1 (ANN2), we detect 163 (177) QSOs with a probability greater than 0.5 and g < 23 in the point-like source catalogue (CL > 0.5). Even though 0.25 is the threshold for an object to be classified as one particular class, we imposed that the probabilities of being a QSO have to be greater than the alternative (P(QSO-h) + P(QSO-l) > P(Galaxy) + P(Star)), which implies P(QSO-h) + P(QSO-l) > 0.5. However, this is a conservative approach because we only considered high-probability objects. If instead we sum all over the probabilities of being QSO, we obtain 182.1 and 195.1 QSOs with the ANN1, and ANN2, respectively.
In Fig. 13, we show the observed (g – r) versus (u – g) colour-colour diagram for all the objects presented in the 1 deg2 mock sample (first row) and in the miniJPAS observations (second row). The positions of QSOs, stars, and galaxies are consistent in each magnitude bin. We included all objects in miniJPAS with FLAGS = 0 and MASK_FLAGS= 0. The 1 deg2 mock sample includes a stellar pouplation that is absend from miniJPAS observations (bottom left side in BIN 0 and 1). These stars correspond to the most massive and bluest stars (O type) that are usually found in regions with a high activity of star formation. The luminosity functions for O and B stars were estimated by extrapolating the prediction of the Besançon model together with the stars within the miniJPAS SDSS superset sample. Therefore, these populations might be overestimated in the 1 deg2 mock sample. However, even if that were the case, the fraction of these stars is still low compared with those on the main sequence. Therefore, the impact that this effect has on a classifier whose main goal is to identify QSO candidates is very limited.
In the last row of Fig. 13, we colour-code the miniJPAS observations with the CL probability. QSOs and stars predicted by the ANN1 are classified by SExtractor as point-like sources (CL > 0.5), while galaxies are predicted as extended sources (CL < 0.5). In the faintest magnitude bin extended and pointlike sources are more difficult to distinguish in the colour space because they both overlap. The ratio of the number of point-like sources according to the CL and the ANN1 (Rpoint = Npoint (CL)/Npoint (ANN1) for BIN 0, 1, and 2 are Rpoint = 0.93, 0.72, and 0.18, respectively, and the ratio of the number of extended sources is Rext = 1.12, 1.06, and 1.49, respectively. We assumed that point-like sources are QSOs and stars, while galaxies were considered extended sources. In summary, our predictions agree with SExtractor considering that the performance of both classifiers decreases as a function of the observed magnitude. Concretely, SExtractor starts to degenerate around 21 magnitude because it is more difficult to distinguish between point-like and extended sources in the low-brightness regime. Furthermore, we predicted based on a sample (miniJPAs observations) that includes extended sources, while our training sample excluded this type of objects.
6 Summary and conclusions
We presented a method based on ANN to classify J spectra into four categories: stars, galaxies, QSO at high redshift (z ≥ 2.1), and QSOs at low redshift (z < 2.1). The algorithms were trained and tested in mock data developed by Queiroz et al. (2022). We employed two different representations of miniJPAS photometry in order to train the algorithms. ANNi used as input photometric fluxes normalised to the detection band (r), while ANN2 employed colours plus the magnitude in the r band. Therefore, ANN1 only has information about the shape of the spectrum, while ANN2 also has access to the observed luminosity.
We enlarged the training set by mixing features from four different classes by adapting the mix up technique. We do not observe significant differences in the performance of the algorithms when a hybrid set was used for the training. A fundamental difference between other works where mix up has been employed with success and this work is probably the complexity of astronomical data. Observations have associated errors that depend at first order on the luminosity of the observed objects. Therefore, features do not encode the same information when objects are brighter or fainter. In other words, mixing between classes appears as a natural outcome in the feature space as the errors increase, which makes faint objects indistinguishable from hybrid bright objects. Thus, hybridisation has an impact on the probabilities yielded by the ANN because they become less realistic when the level of mixing is increased in the training set. Well-calibrated algorithms are as important as obtaining a high performance because the outputs cannot be interpreted as a probability estimation otherwise. In this regard, we showed that the ANNs are better calibrated than the RF classifiers. Furthermore, ANN are better at classifying sources with J-PAS photometry, which was also proven in Rodrigues et al. (2023).
We measured the performance of the algorithms in terms of the f1 score. In the case of quasars at high redshift, we obtain an f1 score with the ANN1 (ANN2) of 0.99 (0.99), 0.93 (0.92), and 0.63 (0.57) for 17 < r ≤ 20, 20 < r ≤ 22.5, and 22.5 <r≤ 23.6, respectively, where r is the J-PAS rSDSS band. For low-redshift quasars, galaxies, and stars, the fl score reached 0.97 (0.97), 0.82 (0.79), and 0.61 (0.58); 0.94 (0.94), 0.90 (0.89), and 0.81 (0.80); and 1.0 (1.0), 0.96 (0.94), and 0.70 (0.52) in the same r bins. We also tested the algorithm on the SDSS test sample, and we obtained a performance compatible with the prediction in the mock test sample. For quasars at high redshift, the fl score with ANN1 (ANN2) reached 1.0 (1.0), and 0.84 (0.86) for 17 < r ≤ 20, and, 20 < r ≤ 22.5. For low-redshift quasars, galaxies, and stars, the fl score reached 0.97 (0.97), and 0.81 (0.82); 0.91 (0.91), and 0.67 (0.73); and 1.0 (1.0), and 0.91 (0.91) in the same r bins. The main source of confusion appears to be between galaxies and low-redshift QSOs. We argue that these two classes are inherently physically mixed, and we provide some examples with SDSS spectra in which the host galaxy of the QSO contributes non-negligibly to the total observed light, showing its dual nature. In these cases, the classifiers mostly yield QSO and galaxy as the two preferred classes. Nevertheless, the SDSS test samples is a relatively small set, and it is only representative of objects brighter than 22.5. Therefore, the results should be treated with caution. The actual performance for fainter objects is still unknown, and the estimates we provided are based on our current physical knowledge of QSO, galaxies, and stars, together with our capability to generate simulated J spectra that mimic miniJPAS observations in the best possible way.
A direct comparison of the results presented in this paper and other works in the literature is challenging because the data are divers and no standardised procedure is available to evaluate the algorithm performance. The capability of distinguishing between QSO, galaxies, and stars clearly depends on many factors, such as the S/N of the observations, the photometric filters, the wavelength coverage, and whether images or integrated photometry are used. For instance, Logan & Fotopoulou (2020) used a clustering algorithm to find star, galaxy, and QSO clusters in a multi-dimensional colour space that included SDSS broad band photometry, Y, H, J, K bands, and WISE mid-infrared observations, together with spatial information such as the half-light radius (HLR). As expected, the performance is a function of the brightness of the source. For bright sources (r < 21), we obtain a similar performance (f1 score ~0.9). Nevertheless, there are not enough objects at magnitudes fainter that 21 for which we could find significant differences. Although the spectral resolution is lower than ours, the spatial information encoded in the HLR and/or the use of infrared colours might improve their classification. He et al. (2021) instead used SDSS images in the u, g, i, and z bands and obtained an f1 score of 0.96, 0.97, and 0.99 in quasars, stars, and galaxies, respectively, for a magnitude brighter than 20. Once again, these results are very similar to ours, but they do not classify faint sources. Nevertheless, the use of images or channels to train convolutions neural networks is very promising as the algorithm can find more complex spatial patterns in the data that might be useful to improve the classification of sources. This approach might be implemented in the future as soon as J-PAS data are available. Finally, it is important to recall that our models were trained with a sample of galaxies and stars that is more likely to be confused with quasars, while the works mentioned above included all type of sources.
In the last section of this paper, we estimated the number of QSOs, galaxies, and stars in the miniJPAS observations, and we showed that our predictions are compatible with previous estimates as well as with SExtractor, which separates between point-like and extended sources. The algorithms presented in this work are part of a combined algorithm that unifies the outcomes of several classifiers (Pérez-Ràfols et al., in prep. b). In the future, we will provide a QSO target list for a spectroscopic follow-up with the WEAVE survey (Jin et al. 2023). This will give us valuable information about the strengths and weakness of our classifiers. WEAVE and the J-PAS collaboration will enter a feedback phase, where the knowledge acquired by one survey is to be transferred to the other in an interactive process. A natural extension of this work when enough spectroscopically confirmed sources are available is to use transfer learning and retraining the algorithms with observations in order to capture the structure of J-PAS data better.
Fig. 6 Fraction of positives for each one of the classes as a function of the mean predicted probability. The ECE for galaxies, QSO-h, QSO-l, stars, and the mean ECE are shown at the top left side of each panel. The error bar represents the standard deviation in each one of the bins. The left panels are trained with the original training set, and the right panel used the hybrid set. The top (bottom) panels show the results of ANN1 (ANN2). |
Fig. 7 Difference between the score obtained with the ANN1 trained in the original training set and the reduced set (ten times smaller), the hybrid set (five time larger), the reduced hybrid set ×5 (five times larger than the reduced set), and the reduced hybrid set ×10 (ten times larger than the reduced set) as a function of the median S/N. Dashed vertical lines indicate an S/N of 10 and 5. |
Fig. 8 score obtained with ANN1 and ANN2 as a function of the median S/N. Dashed vertical lines indicate an S/N of 10 and 5. |
Fig. 9 score and ƒ1 score for each one of the classes obtained within the miniJPAS field observed and labelled according to the SDSS observations (see text in Sect. 4.2). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix are trained with fluxes (see Sect. 3.1). |
Fig. 10 Confusion matrix obtained with ANN1 for the SDSS test sample. |
Fig. 11 Seyfert galaxies observed with miniJPAS and SDSS (see text in Sect. 4.2). The SDSS spectra are scaled to match the miniJPAS r band. The solid grey line represents the actual SDSS observation, and the blue line is a model developed by the SDSS team. We indicate in the legend the miniJPAS ID, the spectroscopic redshift of the object, the class-star yielded by SExtractor (CL), and the AB magnitude in the r band. We also show the probabilities obtained by the ANN1 for each one of the classes, and we attach a multi-colour RGB image centred on the object covering 6.5 arcsec across. All objects except 2470-3341 are classified by the SDSS pipeline as QSOs. |
Fig. 12 Confidence (probability) yielded by the ANN1 (top) and ANN2 (bottom) classifiers for each class and magnitude BIN in miniJPAS observations for point-like sources (CL > 0.5). The numbers of classified objects are shown in the legend. |
Fig. 13 Observed (g – r) vs. (u – g) colour-colour diagram for the 1 deg2 mock sample (first row) and miniJPAS observations (second and third rows). Stars, galaxies, and QSOs are predicted classes with the ANN1 in miniJPAS observations, while they are true classes in the 1deg2 mock sample. The dots in the third row are colour-coded according to the SExtractor probability developed to separate between point-like sources (CL > 0.5) and extended sources (CL < 0.5). Each column includes objects at different magnitude bins. |
Acknowledgements
This paper has gone through internal review by the J-PAS collaboration. G.M.S., R.G.D., R.G.B., L.A.D.G., and J.R.M. acknowledge financial support from the State Agency for Research of the Spanish MCIU through the “Center of Excellence Severo Ochoa” award to the Instituto de Astrofísica de Andalucía (SEV-2017-0709), and to the AYA2016-77846-P and PID2019-109067-GB100. C.Q. acknowledges financial support from the Brazilian funding agencies FAPESP (grants 2015/11442-0 and 2019/06766-1) and Coordenação de Aperfeiçoamento de Pessoal de Nível Su- perior (Capes) – Finance Code 001. I.P.R. was suported by funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowskja-Curie grant agreement no. 754510. M.P.P. and S.S.M. were supported by the Programme National de Cosmologie et Galaxies (PNCG) of CNRS/INSU with INP and IN2P3, co-funded by CEA and CNES, the A*MIDEX project (ANR-11-IDEX-0001-02) funded by the “Investissements d’Avenir” French Government program, managed by the French National Research Agency (ANR), and by ANR under contract ANR-14-ACHN-0021. N.R. acknowledges financial support from CAPES. R.A. was supported by CNPq and FAPESP. J.C.M. and S.B. acknowledge financial support from Spanish Ministry of Science, Innovation, and Universities through the project PGC2018-097585-B-C22. J.A.F.O. acknowledges the financial support from the Spanish Ministry of Science and Innovation and the European Union – NextGenerationEU through the Recovery and Resilience Facility project ICTS-MRR-2021-03-CEFCA. Based on observations made with the JST250 telescope and PathFinder camera for the miniJPAS project at the Observatorio Astrofísico de Javalambre (OAJ), in Teruel, owned, managed, and operated by the Centro de Estudios de Física del Cosmos de Aragón (CEFCA). We acknowledge the OAJ Data Processing and Archiving Unit (UPAD) for reducing and calibrating the OAJ data used in this work. Funding for OAJ, UPAD, and CEFCA has been provided by the Governments of Spain and Aragón through the Fondo de Inversiones de Teruel; the Aragonese Government through the Research Groups E96, E103, E16_17R, and E16_20R; the Spanish Ministry of Science, Innovation and Universities (MCIU/AEI/FEDER, UE) with grant PGC2018-097585-B-C21; the Spanish Ministry of Economy and Competitiveness (MINECO/FEDER, UE) under AYA2015-66211-C2-1-P, AYA2015-66211-C2-2, AYA2012-30789, and ICTS-2009-14; and European FEDER funding (FCDD10-4E-867, FCDD13-4E-2685). Funding for the J-PAS Project has also been provided by the Brazilian agencies FINEP, FAPESP, FAPERJ and by the National Observatory of Brazil with additional funding provided by the Tartu Observatory and by the J-PAS Chinese Astronomical Consortium. We also thank the anonymous referee for many useful comments and suggestions.
Appendix A Confusion matrices and mock test sample
In this section, we show the confusion matrices obtained in the test sample with the ANN1 mix (A.1), ANN2 (A.2), and ANN2 mix (A.3).
Fig. A.1 Confusion matrices obtained with the ANN1 mix in the test sample. |
Fig. A.2 Confusion matrices obtained with the ANN2 in the test sample. |
Fig. A.3 Confusion matrices obtained with the ANN2 mix in the test sample. |
Appendix B Confusion matrices and SDSS test sample
In this section, we show the confusion matrices obtained in the SDSS test sample with the ANN1 mix (B.1), ANN2 (B.2), and ANN2 mix (B.3).
Fig. B.1 Confusion matrix obtained with ANN1 mix in the SDSS test sample. |
Fig. B.2 Confusion matrix obtained with ANN2 in the SDSS test sample. |
Fig. B.3 Confusion matrix obtained with ANN2 mix in the SDSS test sample. |
References
- Abadi, M., Agarwal, A., Barham, P., et al. 2016, ArXiv e-prints [arXiv:1603.04467] [Google Scholar]
- Abramo, L. R., Strauss, M. A., Lima, M., et al. 2012, MNRAS, 423, 3251 [NASA ADS] [CrossRef] [Google Scholar]
- Bai, Y., Liu, J., Wang, S., & Yang, F. 2019, AJ, 157, 9 [Google Scholar]
- Baldry, I. K., Robotham, A. S. G., Hill, D. T., et al. 2010, MNRAS, 404, 86 [NASA ADS] [Google Scholar]
- Baqui, P. O., Marra, V., Casarini, L., et al. 2021, A&A, 645, A87 [EDP Sciences] [Google Scholar]
- Benitez, N., Dupke, R., Moles, M., et al. 2014, ArXiv e-prints [arXiv:1403.5237] [Google Scholar]
- Bishop, C. M. 1995, Neural Networks for Pattern Recognition (USA: Oxford University Press, Inc.) [Google Scholar]
- Bluck, A. F. L., Maiolino, R., Brownson, S., et al. 2022, A&A, 659, A160 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Bonjean, V., Aghanim, N., Salomé, P., et al. 2019, A&A, 622, A137 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Bonoli, S., Marín-Franch, A., Varela, J., et al. 2021, A&A, 653, A31 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Burke, C. J., Aleo, P. D., Chen, Y.-C., et al. 2019, MNRAS, 490, 3952 [NASA ADS] [CrossRef] [Google Scholar]
- Cavuoti, S., Amaro, V., Brescia, M., et al. 2017, MNRAS, 465, 1959 [Google Scholar]
- Chaves-Montero, J., Bonoli, S., Trakhtenbrot, B., et al. 2022, A&A, 660, A95 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Chollet, F., et al. 2015, Keras, https://keras.io [Google Scholar]
- Cooper, M. C., Aird, J. A., Coil, A. L., et al. 2011, ApJS, 193, 14 [NASA ADS] [CrossRef] [Google Scholar]
- Cooper, M. C., Griffith, R. L., Newman, J. A., et al. 2012, MNRAS, 419, 3018 [NASA ADS] [CrossRef] [Google Scholar]
- Croom, S. M., Richards, G. T., Shanks, T., et al. 2009, MNRAS, 399, 1755 [NASA ADS] [CrossRef] [Google Scholar]
- Dalton, G., Trager, S., Abrams, D. C., et al. 2014, SPIE Conf. Ser., 9147, 91470L [Google Scholar]
- Davis, M., Guhathakurta, P., Konidaris, N. P., et al. 2007, ApJ, 660, L1 [NASA ADS] [CrossRef] [Google Scholar]
- Dawson, K. S., Schlegel, D. J., Ahn, C. P., et al. 2013, AJ, 145, 10 [Google Scholar]
- Delli Veneri, M., Cavuoti, S., Brescia, M., Longo, G., & Riccio, G. 2019, MNRAS, 486, 1377 [Google Scholar]
- Dewdney, P. E., Hall, P. J., Schilizzi, R. T., & Lazio, T. J. L. W. 2009, IEEE Proc., 97, 1482 [Google Scholar]
- Domínguez Sánchez, H., Huertas-Company, M., Bernardi, M., Tuccillo, D., & Fischer, J. L. 2018, MNRAS, 476, 3661 [Google Scholar]
- Géron, A. 2019, Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems (O’Reilly Media) [Google Scholar]
- González Delgado, R. M., Díaz-García, L. A., de Amorim, A., et al. 2021, A&A, 649, A79 [Google Scholar]
- González Delgado, R. M., Rodríguez-Martín, J. E., Díaz-García, L. A., et al. 2022, A&A, 666, A84 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Goodfellow, I. J., Shlens, J., & Szegedy, C. 2014, ArXiv e-prints [arXiv: 1412.6572] [Google Scholar]
- Guo, C., Pleiss, G., Sun, Y., & Weinberger, K. Q. 2017, ArXiv e-prints [arXiv:1706.04599] [Google Scholar]
- He, Z., Qiu, B., Luo, A. L., et al. 2021, MNRAS, 508, 2039 [NASA ADS] [CrossRef] [Google Scholar]
- Henrion, M., Mortlock, D. J., Hand, D. J., & Gandy, A. 2011, MNRAS, 412, 2286 [NASA ADS] [CrossRef] [Google Scholar]
- Hernán-Caballero, A., Varela, J., López-Sanjuan, C., et al. 2021, A&A, 654, A101 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Ivezić, Z., Kahn, S. M., Tyson, J. A., et al. 2019, ApJ, 873, 111 [NASA ADS] [CrossRef] [Google Scholar]
- Jin, S., Trager, S. C., Dalton, G. B., et al. 2023, MNRAS, in press [arXiv:2212.03981] [Google Scholar]
- Kim, E. J., & Brunner, R. J. 2017, MNRAS, 464, 4463 [Google Scholar]
- Kim, E. J., Brunner, R. J., & Carrasco Kind, M. 2015, MNRAS, 453, 507 [NASA ADS] [CrossRef] [Google Scholar]
- Krakowski, T., Malek, K., Bilicki, M., et al. 2016, A&A, 596, A39 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Levi, M., Bebek, C., Beers, T., et al. 2013, ArXiv e-prints [arXiv:1308.0847] [Google Scholar]
- Logan, C. H. A., & Fotopoulou, S. 2020, A&A, 633, A154 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- López-Sanjuan, C., Vázquez Ramió, H., Varela, J., et al. 2019, A&A, 622, A177 [Google Scholar]
- Martínez-Solaeche, G., González Delgado, R. M., García-Benito, R., et al. 2021, A&A, 647, A158 [EDP Sciences] [Google Scholar]
- Martínez-Solaeche, G., González Delgado, R. M., García-Benito, R., et al. 2022, A&A, 661, A99 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Nair, V. & Hinton, G. E. 2010, in Proceedings of the 27th International Conference on International Conference on Machine Learning, ICML’10 (Madison, WI, USA: Omnipress), 807 [Google Scholar]
- Oke, J. B., & Gunn, J. E. 1983, ApJ, 266, 713 [NASA ADS] [CrossRef] [Google Scholar]
- Palanque-Delabrouille, N., Magneville, C., Yèche, C., et al. 2016, A&A, 587, A41 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Pâris, I., Petitjean, P., Ross, N. P., et al. 2017, A&A, 597, A79 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Pasquet, J., Bertin, E., Treyer, M., Arnouts, S., & Fouchez, D. 2019, A&A, 621, A26 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Pérez-Ràfols, I., Pieri, M. M., Blomqvist, M., Morrison, S., & Som, D. 2020, MNRAS, 496, 4931 [CrossRef] [Google Scholar]
- Pieri, M. M., Bonoli, S., Chaves-Montero, J., et al. 2016, in SF2A-2016: Proceedings of the Annual meeting of the French Society of Astronomy and Astrophysics, eds. C. Reylé, J. Richard, L. Cambrésy, et al., 259 [Google Scholar]
- Queiroz, C., Abramo, L. R., Rodrigues, N. V. N., et al. 2022, MNRAS, 520, 3476 [Google Scholar]
- Ramachandra, N., Chaves-Montero, J., Alarcon, A., et al. 2022, MNRAS, 515, 1927 [NASA ADS] [CrossRef] [Google Scholar]
- Robin, A. C., Reylé, C., Derrière, S., & Picaud, S. 2003, A&A, 409, 523 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Rodrigues, N. V. N., Raul Abramo, L., Queiroz, C., et al. 2023, MNRAS, 520, 3494 [NASA ADS] [CrossRef] [Google Scholar]
- Rodríguez-Martín, J. E., González Delgado, R. M., Martínez-Solaeche, G., et al. 2022, A&A, 666, A160 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Saglia, R. P., Tonry, J. L., Bender, R., et al. 2012, ApJ, 746, 128 [NASA ADS] [CrossRef] [Google Scholar]
- Shorten, C., & Khoshgoftaar, T. M. 2019, J. Big Data, 6, 60 [CrossRef] [Google Scholar]
- Thulasidasan, S., Chennupati, G., Bilmes, J., Bhattacharya, T., & Michalak, S. 2019, ArXiv e-prints [arXiv:1905.11001] [Google Scholar]
- Vasconcellos, E. C., de Carvalho, R. R., Gal, R. R., et al. 2011, AJ, 141, 189 [Google Scholar]
- Whitten, D. D., Placco, V. M., Beers, T. C., et al. 2019, A&A, 622, A182 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Xiao-Qing, W., & Jin-Meng, Y. 2021, Chinese J. Phys., 69, 303 [NASA ADS] [CrossRef] [Google Scholar]
- Yang, S., Xiao, W., Zhang, M., et al. 2022, ArXiv e-prints [arXiv:2204.08610] [Google Scholar]
- York, D. G., Adelman, J., Anderson, J., John, E., et al. 2000, AJ, 120, 1579 [NASA ADS] [CrossRef] [Google Scholar]
- Zhang, H., Cisse, M., Dauphin, Y. N., & Lopez-Paz, D. 2017, ArXiv e-prints [arXiv:1710.09412] [Google Scholar]
All Tables
All Figures
Fig. 1 score for different magnitude bins as defined in Eq. (14), and the score for the full sample (ALL BIN). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix are trained with fluxes (see Sect. 3.1). |
|
In the text |
Fig. 2 f1 score for each of the classes galaxies, QSO-h, QSO-1, and stars as a function of the magnitude bins defined in Eq. (14), and score for the full sample (ALL BIN). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix do with fluxes (see Sect. 3.1). |
|
In the text |
Fig. 3 Confusion matrices obtained with the ANN1 in the test sample. |
|
In the text |
Fig. 4 Examples of the most typical misclassification of objects in the mock test sample. The first row shows QSO-l classified as galaxies, the second row shows galaxies classified as QSO-l, the third row shows QSO-h classified as QSO-l, and the fourth row shows stars classified as QSO-l. From left to right, the objects are fainter. In each panel, we indicate the AB magnitude in the r band, the redshift (top left), and the probabilities yielded by the ANN1 classifier for each one of the classes (top right). The SDSS spectra from which the test sample is generated are shown in the background in grey. |
|
In the text |
Fig. 5 Median entropy as a function of the median S/N in bins of 1000 objects. |
|
In the text |
Fig. 6 Fraction of positives for each one of the classes as a function of the mean predicted probability. The ECE for galaxies, QSO-h, QSO-l, stars, and the mean ECE are shown at the top left side of each panel. The error bar represents the standard deviation in each one of the bins. The left panels are trained with the original training set, and the right panel used the hybrid set. The top (bottom) panels show the results of ANN1 (ANN2). |
|
In the text |
Fig. 7 Difference between the score obtained with the ANN1 trained in the original training set and the reduced set (ten times smaller), the hybrid set (five time larger), the reduced hybrid set ×5 (five times larger than the reduced set), and the reduced hybrid set ×10 (ten times larger than the reduced set) as a function of the median S/N. Dashed vertical lines indicate an S/N of 10 and 5. |
|
In the text |
Fig. 8 score obtained with ANN1 and ANN2 as a function of the median S/N. Dashed vertical lines indicate an S/N of 10 and 5. |
|
In the text |
Fig. 9 score and ƒ1 score for each one of the classes obtained within the miniJPAS field observed and labelled according to the SDSS observations (see text in Sect. 4.2). Dashed (solid) lines represent the models trained with the hybrid (original) training set. ANN2 and ANN2 mix are trained with colours, while ANN1 and ANN1 mix are trained with fluxes (see Sect. 3.1). |
|
In the text |
Fig. 10 Confusion matrix obtained with ANN1 for the SDSS test sample. |
|
In the text |
Fig. 11 Seyfert galaxies observed with miniJPAS and SDSS (see text in Sect. 4.2). The SDSS spectra are scaled to match the miniJPAS r band. The solid grey line represents the actual SDSS observation, and the blue line is a model developed by the SDSS team. We indicate in the legend the miniJPAS ID, the spectroscopic redshift of the object, the class-star yielded by SExtractor (CL), and the AB magnitude in the r band. We also show the probabilities obtained by the ANN1 for each one of the classes, and we attach a multi-colour RGB image centred on the object covering 6.5 arcsec across. All objects except 2470-3341 are classified by the SDSS pipeline as QSOs. |
|
In the text |
Fig. 12 Confidence (probability) yielded by the ANN1 (top) and ANN2 (bottom) classifiers for each class and magnitude BIN in miniJPAS observations for point-like sources (CL > 0.5). The numbers of classified objects are shown in the legend. |
|
In the text |
Fig. 13 Observed (g – r) vs. (u – g) colour-colour diagram for the 1 deg2 mock sample (first row) and miniJPAS observations (second and third rows). Stars, galaxies, and QSOs are predicted classes with the ANN1 in miniJPAS observations, while they are true classes in the 1deg2 mock sample. The dots in the third row are colour-coded according to the SExtractor probability developed to separate between point-like sources (CL > 0.5) and extended sources (CL < 0.5). Each column includes objects at different magnitude bins. |
|
In the text |
Fig. A.1 Confusion matrices obtained with the ANN1 mix in the test sample. |
|
In the text |
Fig. A.2 Confusion matrices obtained with the ANN2 in the test sample. |
|
In the text |
Fig. A.3 Confusion matrices obtained with the ANN2 mix in the test sample. |
|
In the text |
Fig. B.1 Confusion matrix obtained with ANN1 mix in the SDSS test sample. |
|
In the text |
Fig. B.2 Confusion matrix obtained with ANN2 in the SDSS test sample. |
|
In the text |
Fig. B.3 Confusion matrix obtained with ANN2 mix in the SDSS test sample. |
|
In the text |
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.