Issue |
A&A
Volume 688, August 2024
|
|
---|---|---|
Article Number | A33 | |
Number of page(s) | 12 | |
Section | Catalogs and data | |
DOI | https://doi.org/10.1051/0004-6361/202348714 | |
Published online | 31 July 2024 |
Exploring galactic properties with machine learning
Predicting star formation, stellar mass, and metallicity from photometric data
1
School of Mathematics and Statistics, Xi’an Jiaotong University, Xi’an,
Shaanxi
710049,
PR
China
e-mail: fzeraatgari@xjtu.edu.cn; mosallanezhad@xjtu.edu.cn
2
Department of physics, Institute for Advanced Studies in Basic Sciences,
Zanjan
45195-1159,
Iran
3
CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories,
Beijing,
100101,
PR
China
e-mail: zyx@bao.ac.cn
Received:
23
November
2023
Accepted:
24
May
2024
Aims. We explore machine learning techniques to forecast the star-formation rate, stellar mass, and metallicity across galaxies with redshifts ranging from 0.01 to 0.3.
Methods. Leveraging CatBoost and deep learning architectures, we utilised multiband optical and infrared photometric data from SDSS and AllWISE trained on the SDSS MPA-JHU DR8 catalogue.
Results. Our study demonstrates the potential of machine learning to accurately predict galaxy properties solely from photometric data. We achieved minimised root mean square errors specifically by employing the CatBoost model. For the star-formation rate prediction, we attained a value of RMSESFR = 0.336 dex, while for the stellar mass prediction, the error was reduced to RMSESM = 0.206 dex. Additionally, our model yields a metallicity prediction of RMSEmetallicity = 0.097 dex.
Conclusions. These findings underscore the significance of automated methodologies in efficiently estimating critical galaxy properties amid the exponential growth of multi-wavelength astronomy data. Future research may focus on refining machine learning models and expanding datasets for even more accurate predictions.
Key words: methods: data analysis / methods: statistical / techniques: photometric / astronomical databases: miscellaneous / catalogs / galaxies: star formation
© The Authors 2024
Open Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This article is published in open access under the Subscribe to Open model. Subscribe to A&A to support open access publication.
1 Introduction
The next generation of extensive multi-wavelength photometric sky surveys, exemplified by missions like the Vera C. Rubin Observatory Legacy Survey of Space and Time (LSST; Ivezić et al. 2019), the Euclid Space Telescope (Laureijs et al. 2011), and the China Space Station Telescope (CSST; Zhan 2011), are poised to deliver vast datasets containing critical galaxy properties. These observed properties are not only essential metrics, as they also provide extremely valuable information about various aspects of galaxy evolution, revealing intriguing correlations among them (Brinchmann et al. 2004; Tremonti et al. 2004; Baldry et al. 2008; Lara-Lopez et al. 2010; Mannucci et al. 2010; Kravtsov et al. 2018).
The determination of galaxy properties, such as stellar mass (SM), star-formation rate (SFR), and metallicity, is a complex process that relies on a variety of observational and analytical methods, including the versatile techniques of spectral energy distribution (SED) fitting (Walcher et al. 2011; Conroy 2013). Astronomers use a combination of techniques to uncover the mysteries of galaxies. They utilise telescopes and instruments sensitive to a wide range of wavelengths, spanning from γ ray, X-ray, ultraviolet, and optical bands to infrared and radio bands. These instruments capture diverse emissions from galaxies, revealing the radiance of recently born stars in ultraviolet and optical bands, the infrared signals emitted from dust-covered stellar nurseries, and the faint radiation from old evolved stars within these cosmic structures. Notably, the near-infrared band, with its longer wavelengths, effectively traces the SM linked to the old population, shedding light on the evolution of galaxies (Kennicutt & Evans 2012).
Spectral energy distribution fitting involves assumptions about the past star-formation history of a galaxy, derived from the observed SED and encapsulating a range of astrophysical processes (Ciesla et al. 2017; Bisigello et al. 2016, 2017). Utilising advanced physical models, synthetic SEDs closely resembling the observed data allow for the extraction of critical galaxy properties, including the SFR, SM (M*/SM), age, metallicity, and star-formation history. However, the reliance on pre-existing data, often in the form of template SEDs, introduces inherent biases, influencing the accuracy and interpretation of extracted properties. These biases need careful consideration to ensure the fidelity of derived galaxy characteristics (Conroy 2013; Mitchell et al. 2013; Mobasher et al. 2015; Laigle et al. 2019).
Computational limitations present a challenge, particularly in handling the numerous variables necessary for generating individual spectra. As data projects increase in scale, traditional methods, such as Markov chain Monte Carlo (MCMC) or forward modelling, demand substantial computational resources for assessing properties across a vast number of observed galaxies (Pacifici et al. 2015; Smith & Hayward 2015; Speagle et al. 2016; Jones et al. 2016). This demands more efficient and accurate data analysis, prompting the necessity for innovative approaches such as machine learning (ML)-based statistical techniques.
The advancements in exploring galaxy properties have evolved significantly alongside the rapid progress of ML and deep learning (DL) techniques in recent years. These techniques have demonstrated effectiveness in managing extensive datasets and revealing novel insights (Ball & Brunner 2010; Allen et al. 2019). Machine learning methods not only offer improved computational speed but also the ability to solve complex problems. The recent surge of interest in ML extends across diverse fields in astronomy, impacting various research domains and methodologies, such as star-galaxy classification (Philip et al. 2002; Ball et al. 2006; Vasconcellos et al. 2011; Abraham et al. 2012; Soumagnac et al. 2015; Kim & Brunner 2017; Clarke et al. 2020; Nakoneczny et al. 2021; Cunha & Humphrey 2022; Zeraatgari et al. 2023); galaxy morphology classification (Dieleman et al. 2015; Abraham et al. 2018; Domínguez Sánchez et al. 2018; Barchi et al. 2020; Walmsley et al. 2020; Nair & Vivek 2022); gravitational wave identification (George & Huerta 2018); gravitational lensing identification (Cheng et al. 2020); and estimation of photometric redshifts and other physical properties of galaxies (Tagliaferri et al. 2003; Masters et al. 2015, 2019; Krakowski et al. 2016; Siudek et al. 2018; Hoyle 2016; Stensbo-Smidt et al. 2017; D’Isanto & Polsterer 2018; Turner et al. 2019; Salvato et al. 2019; Bonjean et al. 2019; Delli Veneri et al. 2019; Surana et al. 2020; Mucesh et al. 2021; Razim et al. 2021; Li et al. 2023).
Before exploring how ML and DL impact our research, it is crucial to have a clear understanding of the fundamental distinctions between them. Machine learning typically relies on traditional algorithms and statistical models to learn patterns from data. In traditional ML, the feature engineering process is vital, where domain experts manually select or engineer relevant features for the model. These models are effective for structured data and situations where meaningful insights can be derived from feature engineering, and they perform well with smaller datasets and provide interpretability. This makes traditional ML models suitable for scenarios where understanding the decision-making process of the model is essential (Bishop 2006).
On the other hand, DL is a subset of ML specifically focusing on neural networks with multiple layers. Deep learning models automatically learn hierarchical features from raw data, eliminating the need for extensive feature engineering. They excel at handling large volumes of unstructured data, such as images, and require substantial labelled data for training, often achieving state-of-the-art performance. However, they are considered ‘black-box’ models, as they pose challenges in explaining how they arrive at specific conclusions. Understanding the distinctive characteristics of ML and DL is crucial to maximizing their potential in exploring and comprehending various aspects of galaxies in our research. The fundamental disparity between the two lies in the approach to feature engineering-manual for ML and autonomous hierarchical extraction for DL. This difference shapes the distinct role of each method in influencing the application to comprehend galaxy properties (Goodfellow et al. 2016).
Gaining a comprehensive understanding of galaxy characteristics relies heavily on acquiring optical spectroscopic data, which is often challenging and time-consuming to obtain. Previous studies have acknowledged the complexities involved in accurately determining galaxy properties, especially in establishing the correlation between optical and infrared data and aggregated photometric properties across galaxies (Fogarty et al. 2017; Pearson et al. 2018). In our approach, accurately predicting galaxy properties necessitates seamless integration of both optical and IR data. Specifically, IR data act as a crucial component in estimating properties, serving as a surrogate for SM and portraying the SFR (Leger & Puget 1984). Optical data play a vital role in providing a comprehensive understanding of these properties, as optical spectra enable precise estimations of SFR and corrections for dust attenuation. Additionally, the spectra offer insights into stellar population ages, metallicities, and mass-to-light ratios, while utilising colour-magnitude diagrams is instrumental in determining the total SM and estimating metallicity through key indicators (Kennicutt 1998; Kennicutt & Evans 2012).
Furthermore, our research addresses the challenges of estimating SFR, SM, and metallicity, drawing inspiration from previous studies (Stensbo-Smidt et al. 2017; Bonjean et al. 2019; Delli Veneri et al. 2019; Surana et al. 2020). As these preceding studies have indicated that relying solely on optical data might be insufficient to accurately constrain these galactic properties (Brescia et al. 2014; Stensbo-Smidt et al. 2017; Delli Veneri et al. 2019), our approach, therefore, incorporates both optical and IR data to improve the accuracy and reliability of our results. Additionally, within the current literature, uncertainties emerge due to the absence of a distinct superiority among various ML models, emphasising the lack of reliable benchmarks within the field.
In this investigation, we aim to evaluate the effectiveness of supervised ML, specifically the CatBoost model, and a DL methodology employing the wide and deep neural network (WDNN) architecture. Our approach incorporates optical and IR photometric data to focus on predicting essential galaxy properties–SFR, SM, and metallicity.
Our key objective is to assess the performance and accuracy of these methods in addressing the inherent complexities of estimating these specific galaxy properties. To achieve this, we utilised a combination of magnitudes and colours obtained by cross-matching the SDSS and AllWISE catalogues. In addition, we explored the impact of observational uncertainties by incorporating observational errors in our analysis. Harnessing advanced ML methods, the study focuses on enhancing the accuracy and efficiency of predicting galaxy properties.
This paper is organised as follows. Section 2 provides a detailed description of the data utilised in the study. In Sec. 3, the methodology adopted for analysis and evaluation is outlined. Section 4 covers the ML model and the subsequent results derived from these methodologies, particularly focusing on the performance of ML models for the entire galaxies and various galaxy types while also providing a comparison of our results with other studies. Section 5 summarises our key findings and their implications.
2 Data
The Sloan Digital Sky Survey (SDSS; York et al. 2000) represents one of the most extensive optical surveys in existence, capturing deep images across approximately one-third of the sky in five distinct optical bands: u (λ = 0.355 μm), g (λ = 0.477 μm), r (λ = 0.623 μm), i (λ = 0.762 μm), and z (λ = 0.913 μm). For this study, we used the SQL query tool1 for SDSS to extract optical data from SDSS Data Release 18 (SDSS-DR18).
Our primary focus is on the MPA–JHU DR8 catalogue, a collaborative project between the Max Planck Institute for Astrophysics and Johns Hopkins University (Kauffmann et al. 2003; Brinchmann et al. 2004). This catalogue offers data on SFR, SM, and metallicities for a vast collection of 1 843 200 galaxies, covering a range of redshifts up to approximately z ~ 0.3.
The SFR values (flagged as SFR_TOT_P50) were determined through analysis of Hα emission lines whenever they are present. These values were then adjusted to account for dust extinction using the Balmer decrement Hα/Hβ as detailed in the work by Brinchmann et al. (2004). In the absence of emission lines in certain galaxies, SFRs were instead estimated using a relationship between SFR and the spectral index D4000, as outlined in the studies by Bruzual (1983), Balogh et al. (1999), and Brinchmann et al. (2004). The calculation of SM (flagged as LGM_TOT_P50) relied on theoretical models of stellar populations, following the methodology described by Kauffmann et al. (2003). These computations assumed the application of a Kroupa initial mass function (IMF; Kroupa 2001). Metallicity determination (flagged as OH_P50) was based on the gas-phase oxygen abundance, expressed as 12 + log(O/H), where the ratio O/H signifies the abundance of oxygen relative to hydrogen. Oxygen abundances were only computed for objects classified as ‘star forming’ (Tremonti et al. 2004; Brinchmann et al. 2004).
The Wide-field Infrared Survey Explorer (WISE; Wright et al. 2010) embarked on an all-sky survey project in December 2009, focusing on the mid-infrared spectra with photometry in four filters at W1 (λ = 3.4 μm), W2 (λ = 4.6 μm), W3 (λ = 12 μm), W4 (λ = 22 μm). This endeavour led to the observation of an extensive array of celestial objects, generating over a million images. AllWISE has improved upon WISE by achieving deeper imaging, enhancing source detection, and delivering superior photometric sensitivity and astrometric precision. We adopted three magnitudes that are converted from Vega to AB magnitude as mAB = mVega + Δm, where Δm is 2.699, 3.339, and 5.174 in W1, W2, and W3 bands, respectively (Schindler et al. 2017). Since SDSS filters operate on the AB magnitude system, this conversion ensures that the magnitudes used in our analysis are on the same scale as the SDSS data. In our analysis, we relied on observational errors obtained from SDSS and AllWISE. For SDSS, we utilised the petroMagErr_? values, where ‘?’ corresponds to the five optical magnitudes: u, g, r, i, and z. For AllWISE, we employed the e_W?mag values, where ‘?’ represents the WISE bands W1, W2, and W3.
In order to obtain optical and infrared photometric data, we conducted a cross-match between the SDSS and AllWISE catalogues, using a match radius of 3 arcsec (Su et al. 2013). This process resulted in a dataset comprising 526 409 galaxies. The ML algorithms adopted optical and IR features as inputs and were trained using output data from the SDSS MPA-JHU DR8 catalogue, specifically spectra-extracted SFR, SM, and metallicity. With our study, we aimed at bridging the gaps in understanding and accurately predicting crucial galaxy properties from photometric information.
The SDSS MPA-JHU DR8 catalogue incorporates a classification system denoted by Baldwin–Phillips–Terlevich (BPT) classes (indicated by BPTCLASS; Baldwin et al. 1981). This diagram categorises galaxies based on their emission-line ratios, specifically [O<sc>iii</sc>] /Hβ and [N<sc>ii</sc>] /Hα. Within the classification system of the MPA-JHU catalogue, Class 1 corresponds to star-forming galaxies, Class 2 denotes low signal-to-noise star-forming galaxies, Class 3 represents composite galaxies, and Class 4 is attributed to active galactic nuclei (AGNs), excluding LINERS2. Additionally, Class 5 signifies low signal-to-noise LINERS. The category Class −1 is designed for galaxies unclassified in the BPT diagram, particularly passive galaxies lacking emission lines (Brinchmann et al. 2004).
For our purposes, we selected only data with high quality from the MPA-JHU DR8 catalogue that is considered to be reliable (RELIABLE ≠ 0) and without warning flags in redshift (Z_WARNING = 0). In addition, we kept data points without missing values for total SFR, total SM, and metallic-ity (SFR_TOT_P50 ≠ −9999, LGM_TOT_P50 ≠ −9999, OH_P50 ≠ −9999), and we selected only data with redshift Z > 0.
The dataset sizes are as follows: 129 351 data points for metallicity, 509 711 data points for SFR, and 523 713 data points for SM. The distribution of the data is shown in Fig. 1, from left to right, for SFR, SM, and metallicity.
To better understand the precision of our measurements, we computed the average widths of the 68% confidence intervals for each target parameter across the training dataset. The average width of the 68% confidence interval for the SFR parameter was found to be 0.551 dex. Similarly, for the SM and metallicity parameters, the average widths of the 68% confidence intervals were 0.105 dex and 0.039 dex, respectively.
![]() |
Fig. 1 Distribution of SFR (left panel), SM (middle panel), and metallicity (right panel). |
3 Method
This paper delves into evaluating the performance of two supervised ML models, CatBoost and WDNN, while focusing on quantifying a variety of physical parameters within galaxies. To evaluate the accuracy of estimating these parameters, we undertook a comparative analysis of the two methodologies.
To determine the properties of galaxies using ML based on optical and IR photometric data, we conducted tests involving several models, including random forest, XGBoost, CatBoost, and various DL structures. We explored different neural network structures and architectures, and the final presented architecture emerged from hyperparameter optimisation to ensure a well-tailored and efficient model. After a comprehensive optimisation process, we identified both the WDNN model and CatBoost as the top performers with our dataset. Consequently, our discussions primarily centre around the outcomes derived from these high-performing models.
Given the significantly smaller dataset for metallicity compared to those for SFR and SM, we employed two distinct models-one for predicting SFR and SM and another for predicting metallicity. The dataset size for the ML and WDNN models targeting SFR and SM was 509 198, while the dataset for metallicity encompassed 129 351 instances. We note that the entire prediction process was conducted in logarithmic space for consistency. Initially, to fine-tune the hyperparameters of CatBoost, we split our dataset into dedicated training and test subsets. We randomly allocated 80% for training and reserved 20% for testing in both cases. Subsequently, we applied k-fold cross-validation (Kohavi 1995) with k = 10 exclusively on the training subset for hyperparameter tuning. This method ensures that the test data remains untouched until the final performance evaluation, preserving its independence.
For the neural network, we employed a slightly different strategy. The dataset was initially partitioned into two sets: 20% for testing and 80% for training and hyperparameter tuning. Within the training set, 10% was further allocated for validation, leaving 70% for the actual model training. Here, the validation set served as an independent subset for hyperparameter fine-tuning. Importantly, to ensure better comparability, all results presented in the tables were derived from the same test dataset for both CatBoost and WDNN.
![]() |
Fig. 2 Performance evaluation of CatBoost for different galaxy property predictions with a varying ‘depth’ based on the mean absolute error for SFR (left panel), SM (middle panel), and metallicity (right panel). The results presented are the average performance across a ten-fold cross-validation. |
3.1 CatBoost
CatBoost, a member of the gradient-boosted decision tree (GBDT) family introduced by Friedman (2001) and developed by Dorogush et al. (2018), is a high-performance open-source algorithm that utilises boosting techniques to construct resilient predictive models. In gradient boosting, an ensemble of weak learners, typically decision trees, is built sequentially. Each tree corrects the errors of the preceding ones, gradually improving the overall predictive performance. Specifically, the algorithm focuses on minimising the residuals of the previous trees, enhancing the ability of the model to capture complex relationships in the data.
Moreover, CatBoost incorporates regularisation techniques to prevent overfitting and enhance model generalizability. The algorithm also provides built-in functionalities for cross-validation, streamlining hyperparameter tuning and facilitating model performance assessment. The central hyperparameter, ‘depth’, plays a pivotal role in controlling the complexity of the decision trees. It determines the maximum depth of each tree in the ensemble. A higher depth allows the model to capture intricate patterns in the training data, but it also increases the risk of overfitting. On the other hand, a lower depth may lead to underfitting, where the model fails to capture essential details. Striking the right balance for the depth parameter is crucial to achieving optimal predictive performance. Notably, the default settings of CatBoost often yield a strong performance, with optimisation efforts typically focused on adjusting the depth parameter while retaining other parameters at their default values.
In our study, ten-fold cross-validation was conducted, and mean absolute error (MAE) values with varying ‘depth’ are illustrated in Fig. 2. Optimal performance was achieved with a depth parameter set to ten, resulting in MAE values of about 0.259 dex, 0.147 dex, and 0.082 dex for SFR, SM, and metallicity, respectively.
3.2 Deep learning
In our study, we employed the wide and deep model (Cheng et al. 2016) for DL using Keras (Chollet et al. 2015). Unlike the conventional neural networks that require data to pass through each layer, potentially obscuring simpler patterns, WDNN allows for non-sequential processing. This capability enables the network to simultaneously comprehend complex patterns (via the deep path) and simple rules (via the short path). This unique design fosters a more comprehensive understanding of the complexities of the data compared to the sequential processing of traditional neural networks.
The architecture of the two WDNNs used in our study is shown in Fig. 3. The model displayed in the left panel aims to predict SFR and SM, while the model in the right panel focuses solely on predicting metallicity. The left model consists of six hidden layers, each structured with neurons arranged as [16, 16, 32, 32, 64, 64] and utilising rectified linear unit (ReLU) activation functions. A concatenate layer combines the input and output of the sixth hidden layer, resulting in the output layer comprising two neurons (SFR and SM) with a linear activation function.
The right model in Fig. 3, tailored towards the prediction of metallicity, entails three hidden layers with [16, 32, 64] neurons and uses the exponential linear unit (ELU) activation function. The third hidden layer of this model is then concatenated with the input and output of the concatenate layer. The output layer here consists of a single neuron (metallicity) with a linear activation function.
Different from the CatBoost model, WDNN involves refining various hyperparameters, such as the number of neurons, layers, activation functions, and kernel initialisers. These hyperparameters undergo careful adjustments during training, spanning up to 1000 epochs with the Adam optimiser, and the training process employs the mean square error (MSE) as the loss function. Training continues until the difference between loss values of two consecutive epochs is below the machine precision level.
Furthermore, we integrated the ReduceLROnPlateau technique, which dynamically modifies the learning rate based on the performance of the model on a dedicated validation set. Specifically, if the performance metric remains constant for several consecutive epochs during training, we systematically decreased the learning rate by a specified amount. This systematic adjustment not only improved model performance but also enhanced the model’s adaptability to the complexities of the data. Lowering the learning rate helped the model navigate the optimisation landscape more effectively, potentially aiding in overcoming local minima and expediting convergence or fine-tuning.
![]() |
Fig. 3 Schematic structure of the WDNN models utilised in this study. The input features consist of a combination of eight optical and IR magnitudes and colours. The left panel outputs SFR and SM, while the right panel focuses on predicting metallicity. |
3.3 Evaluation metrics
We evaluated the accuracy and efficacy of our regression models using metrics such as the residual between the true values (targets, Xt) of the galaxy properties obtained by SED fitting through spectra and the predicted values (Xp) by our regression models for each property in our dataset as
(1)
where X represents any of the three galaxy properties (SFR, SM, or metallicity). These differences serve as quantitative indicators of how well our models approximate the true galaxy properties.
The MSE measures the average squared difference between the predicted values and the true values as
(2)
where Xt,i is the true value for the ith galaxy, Xp,i is the predicted value for the ith galaxy, and n is the sample size. The root mean squared error (RMSE) is the square root of the MSE and provides an interpretable scale similar to the target variable as
(3)
The normalised root mean squared error (NRMSE) is a normalised version of the RMSE, and it was calculated by dividing the RMSE by the range (the difference between the maximum and minimum values) of the true target values:
(4)
The mean absolute error (MAE) measures the average absolute difference between the predicted values and the true values. It is less sensitive to outliers than MSE.
(5)
The standard deviation of the residuals was calculated as
(6)
Bias is defined as the average separation between prediction and true values:
(7)
In our study on estimating galaxy properties, η represents the percentage of catastrophic outliers. We applied a criterion of | ΔX |> 3σ, where data points with residuals exceeding three times the standard deviation are considered potential catastrophic outliers.
4 Machine learning model
4.1 Feature selection
The careful choice of input features plays an important role in achieving the optimal performance of an ML model. In our experiments, a set of eight optical and IR magnitudes and colours, including r, u − g, g − r, r − i, i − z, z − W1, W1 − W2, and W2 − W3, from previous studies (Bonjean et al. 2019; Li et al. 2023) was employed as input features for the training of the models.
The importance of a feature in addressing a learning problem is evaluated based on its contribution. By employing CatBoost, we assessed the importance scores for each feature and organised them in descending order based on their significance, as depicted in Fig. 4. The CatBoost importance score, derived from this evaluation, quantifies the contribution of each feature to the predictive performance of the model. A higher CatBoost importance score signifies a more significant role of a feature in predicting the target variable within the model. Importantly, these feature-importance scores provided by CatBoost consider not only linear relationships but also non-linear contributions of features to the predictive capabilities of the model. As is evident from Fig. 4, the most influential features for predicting SFR and SM are IR colours W2 − W3 and W1 − W2, with scores of 18.37% and 16.05%, respectively. In contrast, for predicting metallicity, the key features are both optical and IR colours g − r and W2 − W3, with scores of 18.01% and 17.43%, respectively.
![]() |
Fig. 4 Feature importance analysis by CatBoost. The left panel demonstrates the significance of features for predicting SFR and SM, while the right panel provides insights into feature importance for predicting metallicity. |
Comparative evaluation metrics of ML models (CatBoost and WDNN) for the physical parameters SFR, SM, and metallicity.
4.2 Results
4.2.1 Performance of CatBoost and WDNN
We trained the CatBoost and WDNN models on the training set to estimate the SFR, SM, and metallicity of the test set. The assessment of the models was performed by comparing the results with the true values from the SDSS MPA-JHU DR8 catalogue. For a comprehensive evaluation, we employed a range of key metrics outlined in Table 1, including RMSE, NRMSE, MAE, bias, and the proportion of catastrophic outliers η.
Our methodology involved a dual-model approach to evaluate CatBoost and WDNN. Each model was examined in two separate configurations, as each focused on different output targets. The first model predicts SFR and SM, while the second model predicts metallicity. The upper row subplots in Figs. 5, 6, and 7 represent the disparities between the ML estimated values and the true data. Below these subplots, the errors are plotted about the true values, providing a visual representation of how the residuals fluctuate concerning the actual data points. The consistent performance of both ML models is clearly demonstrated by the strong agreement between the true values and predictions by both ML algorithms. When comparing the performance of CatBoost (left panels) and WDNN (right panels), we found a comparable result.
The results presented in Table 1 further reinforce that the performances of CatBoost and WDNN are comparable across all metrics. Minimal differences, however, were observed, with CatBoost generally demonstrating slightly lower RMSE values than the WDNN model, indicating marginal advantages, typically around 0.001.
The NRMSEs indicate that the models exhibit high accuracy in predicting galaxy parameters, as evidenced by their small values. Specifically, the NRMSE for SFR is 0.040; for SM, it is 0.034; and for metallicity, it is 0.064. Analysing these values, we observed that the highest NRMSE is associated with metallicity. Therefore, it can be inferred that metallicity poses a greater challenge for prediction compared to the other two parameters. This could imply that the features or factors influencing metallicity in galaxies are more complex or less straightforward for the model to capture accurately. It may also indicate that there is greater variability or uncertainty in the data related to metallicity. Additionally, the comparatively worse performance of the metallicity prediction could be related to the fact that the training data for these models is about a factor of five smaller than that for SFR and SM.
The bottom row subplots of Figs. 5, 6, and 7 illustrate a comparison of model accuracy across different redshifts. These figures present both residual scatter plots and box plots plotted alongside redshift. Upon examination, we observed symmetrical box plots, indicating that the variability, or spread, in the residuals is consistent across the range of redshifts being considered. However, we also noted that at lower redshifts, the whisker lengths in the box plots were significantly longer compared to higher redshifts. This observation hints at a greater variability, or spread, in the residuals at lower redshifts, potentially posing challenges for accurate prediction of galaxy properties within this range. Furthermore, we observed a more pronounced distance between the median of the box plot and the zero residual line at low redshifts compared to high redshifts. This observation implies larger deviations from zero residuals on average within this range, which may further complicate accurate predictions in lower redshift regions.
![]() |
Fig. 5 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of SFR estimates from the ML model with SFR values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the SFR errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
4.2.2 Enhanced predictive analysis: Exploring CatBoost with error considerations
Our ML models were trained using observational data characterised via spectroscopy, a valid approach that is common in such studies (Bonjean et al. 2019; Delli Veneri et al. 2019; Surana et al. 2020). However, it is crucial to acknowledge that the accuracy of our presented models is inherently tied to the quality of the training data. Therefore, we have incorporated measurement uncertainties from the training data during the training process. This ensures that our models not only learn from the nominal values but also consider the potential variations and uncertainties in the data, contributing to a more robust and reliable predictive framework.
In the previous section, we concentrated on using magnitudes and colours as features in our predictive models. Now, we broaden our approach to consider the observational errors in our modelling process. We opted for CatBoost for this experiment because it showed a better performance than WDNN. Our main goal here is to thoroughly evaluate how magnitude uncertainties affect the predictive accuracy of our model.
To conduct this experiment, we removed entries with null errors and those where the errors exceeded one. This reduced the available training data to 364 424 data points for the SFR and SM prediction and to 121 060 data points for the metallicity prediction. We re-evaluated our pre-trained CatBoost models from Sect. 4.2.1 using this updated dataset that now includes observational errors. In contrast, for the models with errors, we developed new ML models trained specifically on this refined dataset. Following a similar approach to the one in the previous section, we conducted hyperparameter tuning on CatBoost and then used ten-fold cross-validation exclusively on the training data. Through this process, we identified that a depth parameter of ten results in optimal performance.
In Table 2, we present the outcomes of this analysis and compare the performance of the CatBoost model with and without errors on the same dataset. Our results show that incorporating observational errors into the models yields enhanced performance compared to models that overlook such uncertainties. Specifically, the RMSE for SFR decreases to 0.336 dex when errors are considered, compared to 0.353 dex without error considerations. Similarly, when examining the performance metrics with and without errors, we observed a reduction in RMSE for the SM prediction, dropping from 0.216 dex to 0.206 dex. Likewise, for the metallicity prediction, the RMSE decreases from 0.104 dex to 0.097 dex.
![]() |
Fig. 6 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of SM estimates from the ML model with SM values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the SM errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
![]() |
Fig. 7 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of metallicity estimates from the ML model with metallicity values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the metallicity errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
Comparison of CatBoost model performance on physical parameters with and without observational errors in input features but using the same dataset.
4.2.3 Performance of CatBoost for different kinds of galaxies
In the preceding subsections, we conducted predictions without regard to galaxy type. This approach facilitated a comprehensive assessment of the predictive capabilities of our ML models and allowed us to evaluate their generalizability across a diverse array of galaxies.
To characterise the performance of our trained models in more detail, we carried out focused analyses for each specific group of galaxies. Given the better performance of CatBoost compared to the WDNN, we exclusively employed the CatBoost model, which accounts for observational errors, for this analysis. Thus, we evaluated the predictive performance of the CatBoost algorithm based on the BPT classifications outlined in the SDSS MPA-JHU catalogue. Our test dataset was categorised into four groups: passive (14 835 data points), AGN (11 903 data points), composite (6600 data points), and star forming (41 869 data points). As the BPT classifications for galaxies with available metallicities were absent in the SDSS MPA-JHU catalogue, we did not analyse metallicity in this case.
We display the results of this analysis in Fig. 8, highlighting the difference between the predicted values of SFR (upper row subplots) and SM (bottom row subplots) generated by CatBoost and the true values from the SDSS MPA-JHU catalogue. The comparisons were conducted across various BPT classes, organised from left to right as passive, AGNs, composites, and star-forming galaxies. The robust consistency between the true values and the predicted values is evident across various galaxy categories.
Table 3 provides a detailed overview of the evaluation metrics, encompassing RMSE, NRMSE, MAE, bias, and η specifically across galaxy types. We observed an overall good performance across all types of galaxies, but a closer analysis of the results revealed significant variations in predictive accuracy, especially in the transition from passive to more active galaxies. This transition significantly impacts both the SFR and SM predictions.
The NRMSE values for the SFR prediction indicate that star-forming galaxies exhibit the lowest NRMSE, at 0.040, whereas passive galaxies are at 0.045, composite galaxies at 0.065, and AGNs at 0.092. Likewise, for the SM prediction, passive galaxies demonstrate the most accurate prediction, with an NRMSE of 0.025, followed by star-forming galaxies at 0.036, AGN galaxies at 0.037, and composite galaxies at 0.052. The higher evaluation metrics in AGNs, composites, and star-forming galaxies suggest increased prediction errors for both SFR and SM compared to passive galaxies.
![]() |
Fig. 8 Results of CatBoost in predicting SFR and SM for different kinds of galaxies. |
Performance of CatBoost model for different kinds of galaxies.
4.2.4 Comparison to other works
In order to make comparisons with other works, it is crucial to acknowledge at the outset the inherent variations among studies arising from the utilisation of different data and input patterns. Table 4 only provides a simple comparative analysis of methodologies employed in the automated prediction of galaxy properties (as explored by Bonjean et al. 2019; Delli Veneri et al. 2019; Surana et al. 2020). Additionally, we include our best-performing model, the CatBoost model that accounts for observational errors, in this table for comparison. Bonjean et al. (2019) conducted a comprehensive study using the SDSS catalogue and cross-matched it with the AllWISE catalogue. Their input features comprised IR magnitudes and spectroscopic redshifts, and they employed a random forest algorithm trained on SFR and SM data extracted from the SDSS MPA-JHU catalogue. Since the RMSE and σ values in our results are identical (because the mean of ΔXi in Eq. (6) effectively vanishes), we specify that the σ reported by Bonjean has been incorporated into the RMSE column of this table. Their reported errors were σSFR = 0.38 dex and σSM = 0.16 dex, whereas our model yielded 0.336 dex and 0.206 dex, respectively. Delli Veneri et al. (2019) used 55 photometric features (magnitudes, colours, and photoz). Their method involved estimating specific SFR (sSFR = SFR/M*) using such techniques as parameter space optimisation and employing a supervised ML model, such as a random forest. Their reported error for the specific SFR was RMSEsSFR = 0.252 dex. Compared to these two works, we applied eight optical and infrared magnitudes and colours (r, u − g, g − r, r − i, i − z, z − W1, W1 − W2, W2 − W3), and only based on photometric information without spectroscopic redshifts, we arrived at an RMSEsSFR of 0.248 dex.
Surana et al. (2020) utilised data from the Galaxy And Mass Assembly (GAMA) survey, which incorporates a wide range of input features, such as from far-UV, near-UV, optical, near-IR, and mid-IR bands and spectroscopic redshifts. They employed DL techniques to predict three properties of galaxies. Two of our galaxy properties align with two of theirs, namely, SFR and SM. Comparing the results of predicting these two parameters, the reported errors for their predictions were that RMSESFR = 0.1658 dex and RMSESM = 0.0582 dex. In comparison, our model produced errors of 0.336 dex and 0.206 dex for the same properties, respectively. This suggests that, as indicated by Surana et al. (2020), having additional information from multiple bands, specifically from UV, optical, and IR, enhances the accuracy of estimating physical parameters.
Model performance comparison with related studies.
5 Summary and conclusions
In this research, we have explored the application of ML models by specifically utilising CatBoost and WDNN to predict essential galaxy properties – SFR, SM, and metallicity. Our approaches were trained using optical and IR photometric data from SDSS DR18 and AllWISE, with the SDSS MPA-JHU DR8 values serving as a target for training. In addition, we incorporated observational errors into our models to offer a comprehensive assessment of the accuracy achievable with the trained ML methods.
In our methodology, we adopted a dual-model strategy for both of the ML techniques under investigation. The first model was configured to accommodate two outputs, specifically the prediction of SFR and SM. For the second model, our focus was on training it with a singular output, specifically predicting metallicity. Moreover, we introduced innovative DL architectures, identifying two effective WDNN structures that achieved competitive results compared to prior studies.
Notably, during this analysis, we found that CatBoost outperformed WDNN. To further explore the impact of observational uncertainties on our predictions, we conducted an additional experiment incorporating observational errors alongside magnitudes and colours and using CatBoost exclusively. The results indicated a significantly improved performance when observational errors were considered. This improvement was particularly evident across all three galaxy parameters compared to when the errors were not factored in. This finding suggests the importance of addressing observational data uncertainties in order to achieve improved predictive accuracy.
Subsequently, we conducted a thorough analysis of various galaxy types categorised by the BPT classifications in the same catalogue. The galaxies were divided into passive and active groups, encompassing AGN, composite, and star-forming categories. In this test, our CatBoost model that accounts for observational errors demonstrated its robustness, providing a reliable performance across all considered galaxy types. We did, however, observe a decrease in predictive performance for the more active galaxies.
In conclusion, our study underscores the adaptability of both the CatBoost and WDNN models in predicting galaxy properties using optical and IR photometric data. The expanding volume of astronomical data from larger surveys has created a growing demand for accurate and timely predictions, making ML a particularly suitable solution. Our models demonstrate the capability to provide predictions with a high degree of certainty, offering promising avenues for further exploration in ongoing and upcoming photometric surveys, such as by the James Webb Space Telescope (JWST), LSST, and Euclid. Opportunities for improvement of prediction accuracy and data completeness lie in integrating information derived from broad surveys with comprehensive observations.
Acknowledgements
We specially thank the reviewer for his invaluable feedback and insightful comments to help us improve our paper. This paper is funded by the National Natural Science Foundation of China under grant no. 12273076 (Y.Z.) and the science research grants from the China Manned Space Project with nos. CMS-CSST-2021-A04 and CMS-CSST-2021-A06. We acknowledge the use of SDSS and WISE databases. Funding for the Sloan Digital Sky Survey IV has been provided by the Alfred P. Sloan Foundation, the U.S. Department of Energy Office of Science, and the Participating Institutions. SDSS-IV acknowledges support and resources from the Centre for High-Performance Computing at the University of Utah. The SDSS website is www.sdss.org. Funding for the SDSS has been provided by the Alfred P. Sloan Foundation, the Participating Institutions, the National Science Foundation, the U.S. Department of Energy, the National Aeronautics and Space Administration, the Japanese Monbuka-gakusho, the Max Planck Society, and the Higher Education Funding Council for England. The SDSS Web Site is http://www.sdss.org/. The SDSS is managed by the Astrophysical Research Consortium for the Participating Institutions. The Participating Institutions are the American Museum of Natural History, Astrophysical Institute Potsdam, University of Basel, University of Cambridge, Case Western Reserve University, University of Chicago, Drexel University, Fermilab, the Institute for Advanced Study, the Japan Participation Group, Johns Hopkins University, the Joint Institute for Nuclear Astrophysics, the Kavli Institute for Particle Astrophysics and Cosmology, the Korean Scientist Group, the Chinese Academy of Sciences (LAMOST), Los Alamos National Laboratory, the Max-Planck-Institute for Astronomy (MPIA), the MaxPlanck-Institute for Astrophysics (MPA), New Mexico State University, Ohio State University, University of Pittsburgh, University of Portsmouth, Princeton University, the United States Naval Observatory, and the University of Washington. This publication makes use of data products from the Wide-field Infrared Survey Explorer, which is a joint project of the University of California, Los Angeles, and the Jet Propulsion Laboratory/California Institute of Technology, funded by the National Aeronautics and Space Administration.
References
- Abraham, S., Philip, N. S., Kembhavi, A., Wadadekar, Y. G., & Sinha, R. 2012, MNRAS, 419, 80 [NASA ADS] [CrossRef] [Google Scholar]
- Abraham, S., Aniyan, A. K., Kembhavi, A. K., Philip, N. S., & Vaghmare, K. 2018, MNRAS, 477, 894 [NASA ADS] [CrossRef] [Google Scholar]
- Allen, G., Andreoni, I., Bachelet, E., et al. 2019, arXiv e-prints [arXiv:1982.88522] [Google Scholar]
- Baldry, I. K., Glazebrook, K., & Driver, S. P. 2008, MNRAS, 388, 945 [NASA ADS] [Google Scholar]
- Baldwin, J. A., Phillips, M. M., & Terlevich, R. 1981, PASP, 93, 5 [Google Scholar]
- Ball, N. M., & Brunner, R. J. 2010, Int. J. Mod. Phys. D, 19, 1049 [Google Scholar]
- Ball, N. M., Brunner, R. J., Myers, A. D., & Tcheng, D. 2006, ApJ, 650, 497 [NASA ADS] [CrossRef] [Google Scholar]
- Balogh, M. L., Morris, S. L., Yee, H. K. C., Carlberg, R. G., & Ellingson, E. 1999, ApJ, 527, 54 [Google Scholar]
- Barchi, P. H., de Carvalho, R., Rosa, R., et al. 2020, Astron. Comput., 30, 100334 [NASA ADS] [CrossRef] [Google Scholar]
- Bishop, C. M. 2006, Pattern Recognition and Machine Learning (Springer) [Google Scholar]
- Bisigello, L., Caputi, K. I., Colina, L., et al. 2016, ApJS, 227, 19 [NASA ADS] [CrossRef] [Google Scholar]
- Bisigello L., Caputi, K. I., Colina, L., et al. 2017, ApJS, 231, 3 [NASA ADS] [CrossRef] [Google Scholar]
- Bonjean, V., Aghanim, N., Salomé, P., et al. 2019, A&A, 622, A137 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Brescia, M., Cavuoti, S., Longo, G., & De Stefano, V. 2014, A&A, 568, A126 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Brinchmann, J., Charlot, S., White, S. D. M., et al. 2004, MNRAS, 351, 1151 [Google Scholar]
- Bruzual, A. G. 1983, ApJ, 273, 105 [CrossRef] [Google Scholar]
- Bruzual, G., & Charlot, S. 2003, MNRAS, 344, 1000 [NASA ADS] [CrossRef] [Google Scholar]
- Calzetti, D., Kinney, A. L., & Storchi-Bergmann, T. 1994, ApJ, 429, 582 [Google Scholar]
- Cheng, H.-T., et al. 2016, “Wide & Deep Learning for Recommender Systems,” Proceedings of the First Workshop on Deep Learning for Recommender Systems, 7 [CrossRef] [Google Scholar]
- Cheng, T.-Y., Li, N., Conselice, C. J., et al. 2020, MNRAS, 494, 3750 [NASA ADS] [CrossRef] [Google Scholar]
- Chollet, F., et al. 2015, Keras, https://keras.io [Google Scholar]
- Ciesla, L., Elbaz, D., & Fensch, J. 2017, A&A, 608, A41 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Clarke, A. O., Scaife, A. M. M., Greenhalgh, R., & Griguta, V. 2020, A&A, 639, A84 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Conroy, C. 2013, ARA&A, 51, 393 [NASA ADS] [CrossRef] [Google Scholar]
- Cunha, P. A. C., & Humphrey, A. 2022, A&A, 666, A87 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Delli Veneri, M., Cavuoti, S., Brescia, M., Longo, G., & Riccio, G. 2019, MNRAS, 486, 1377 [Google Scholar]
- Dieleman, S., Willett, K. W., & Dambre, J. 2015, MNRAS, 450, 1441 [NASA ADS] [CrossRef] [Google Scholar]
- D’Isanto, A., & Polsterer, K. L. 2018, A&A, 609, A111 [Google Scholar]
- Domínguez Sánchez, H., Huertas-Company, M., Bernardi, M., Tuccillo, D., & Fischer, J. L. 2018, MNRAS, 476, 3661 [Google Scholar]
- Dorogush, A. V., Ershov, V., & Yandex, A. G. 2018, arXiv e-prints [arXiv:1810.11363] [Google Scholar]
- Dressler, A., Lynden-Bell, D., Burstein, D., et al. 1987, ApJ, 313, 42 [Google Scholar]
- Faber, S. M. 1973, ApJ, 179, 731 [NASA ADS] [CrossRef] [Google Scholar]
- Fogarty, K., Postman, M., Larson, R., Donahue, M., & Moustakas, J. 2017, ApJ, 846, 103 [Google Scholar]
- Friedman, J. H. 2001, Ann. Stat., 29, 1189 [Google Scholar]
- Gallazzi, A., Charlot, S., Brinchmann, J., White, S. D. M., & Tremonti, C. A. 2005, MNRAS, 362, 41 [Google Scholar]
- Garnett, D. R. 2002, ApJ, 581, 1019 [NASA ADS] [CrossRef] [Google Scholar]
- George, D., & Huerta, E. A. 2018, Phys. Lett. B, 778, 64 [NASA ADS] [CrossRef] [Google Scholar]
- Goodfellow, I., Bengio, Y., & Courville, A. 2016, Deep Learning (Cambridge: MIT Press) [Google Scholar]
- Hoyle, B. 2016, Astron. Comput., 16, 34 [NASA ADS] [CrossRef] [Google Scholar]
- Ivezić, Ž., et al., 2019, ApJ, 873, 111 [NASA ADS] [CrossRef] [Google Scholar]
- Janowiecki, S., Catinella, B., Cortese, L., et al., 2017, MNRAS, 466, 4795 [NASA ADS] [Google Scholar]
- Jones, M. L., Hickox, R. C., Black, C. S., et al. 2016, ApJ, 826, 12 [NASA ADS] [CrossRef] [Google Scholar]
- Kauffmann, G., Heckman, T. M., White, S. D. M., et al. 2003, MNRAS, 341, 33 [Google Scholar]
- Kennicutt, Jr. R. C. 1998, ApJ, 498, 541 [NASA ADS] [CrossRef] [Google Scholar]
- Kennicutt, R. C., & Evans, N. J. 2012, ARA&A, 50, 531 [NASA ADS] [CrossRef] [Google Scholar]
- Kim, E. J., & Brunner, R. J. 2017, MNRAS, 464, 4463 [Google Scholar]
- Kohavi, R. 1995, Proceedings of the 14th International Joint Conference on Artificial Intelligence – Volume 2, IJCAI’95 [Google Scholar]
- Kormendy, J. 1977, ApJ, 218, 333 [NASA ADS] [CrossRef] [Google Scholar]
- Krakowski, T., Małek, K., Bilicki, M., et al. 2016, A&A, 596, A39 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Kravtsov, A. V., Vikhlinin, A. A., & Meshcheryakov, A. V. 2018, Astron. Lett., 44, 8 [Google Scholar]
- Kroupa, P. 2001, MNRAS, 322, 231 [NASA ADS] [CrossRef] [Google Scholar]
- Lagache, G., Puget, J.-L., & Dole, H. 2005, ARA&A, 43, 727 [NASA ADS] [CrossRef] [Google Scholar]
- Laigle, C., Davidzon, I., Ilbert, O., et al. 2019, MNRAS, 486, 5104 [Google Scholar]
- Lara-Lopez, M. A., Cepa, J., Bongiovanni, A., et al. 2010, A&A, 521, A53 [Google Scholar]
- Laureijs, R., et al., 2011, arXiv e-prints [arXiv:1110.3193] [Google Scholar]
- Leger, A., & Puget, J. L. 1984, A&A, 137, L5 [Google Scholar]
- Lequeux, J., Peimbert, M., Rayo, J. F., Serrano, A., & Torres-Peimbert S. 1979, A&A, 80, 155 [NASA ADS] [Google Scholar]
- Li, C., Zhang, Y., Chenzhou, C., et al. 2023, MNRAS, 518, 513 [Google Scholar]
- Mannucci, F., Cresci, G., Maiolino, R., Marconi, A., & Gnerucci, A. 2010, MNRAS, 408, 2115 [NASA ADS] [CrossRef] [Google Scholar]
- Masters, D., Capak, P., Stern, D., et al. 2015, ApJ, 813, 53 [Google Scholar]
- Masters, D. C., Stern, D. K., Cohen, J. G., et al. 2019, ApJ, 877, 81 [Google Scholar]
- Mitchell, P. D., Lacey, C. G., Baugh, C. M., & Cole, S. 2013, MNRAS, 435, 87 [NASA ADS] [CrossRef] [Google Scholar]
- Mobasher, B., Dahlen, T., Ferguson, H. C., et al. 2015, ApJ, 808, 101 [NASA ADS] [CrossRef] [Google Scholar]
- Mucesh, S., Hartley, W. G., Palmese, A., et al. 2021, MNRAS, 502, 2770 [NASA ADS] [CrossRef] [Google Scholar]
- Nair, A., & Vivek, M. 2022, MNRAS, 511, 4946 [NASA ADS] [CrossRef] [Google Scholar]
- Nakoneczny, S. J., Bilicki, M., Pollo, A., et al. 2021, A&A, 649, A81 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Pacifici, C., da Cunha, E., Charlot, S., et al. 2015, MNRAS, 447, 786 [NASA ADS] [CrossRef] [Google Scholar]
- Pearson, W. J., Wang, L., Hurley, P. D., et al. 2018, A&A, 615, A146 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Philip, N. S., Wadadekar, Y., Kembhavi, A., & Joseph, K. B. 2002, A&A, 385, 1119 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Razim, O., Cavuoti, S., Brescia, M., et al. 2021, MNRAS, 507, 5034 [NASA ADS] [CrossRef] [Google Scholar]
- Roberts, M. S., & Haynes, M. P. 1994, ARA&A, 32, 115 [NASA ADS] [CrossRef] [Google Scholar]
- Salim, S., Rich, R. M., Charlot, S., et al. 2007, ApJS, 173, 267 [NASA ADS] [CrossRef] [Google Scholar]
- Salvato, M., Ilbert, O., & Hoyle, B. 2019, Nat. Astron. 3, 212 [NASA ADS] [CrossRef] [Google Scholar]
- Schindler, J. T., Fan, X., McGreer, I. D., et al. 2017, ApJ, 851, 13 [NASA ADS] [CrossRef] [Google Scholar]
- Siudek, M., Małek, K., Pollo, A. et al. 2018, A&A, 617, A70 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
- Smith, D. J. B., & Hayward, C. C. 2015, MNRAS, 453, 1597 [NASA ADS] [CrossRef] [Google Scholar]
- Soumagnac, M. T., Abdalla, F. B., Lahav, O., et al. 2015, MNRAS, 450, 666 [NASA ADS] [CrossRef] [Google Scholar]
- Speagle, J. S., Capak, P. L., Eisenstein, D. J., Masters, D. C., & Steinhardt, C. L. 2016, MNRAS, 461, 3432 [NASA ADS] [CrossRef] [Google Scholar]
- Stensbo-Smidt, K., Gieseke, F., Igel, C., Zirm, A., & Steenstrup Pedersen, K. 2017, MNRAS, 464, 2577 [NASA ADS] [CrossRef] [Google Scholar]
- Su, S., Kong, X., Li, J., & Fang, G. 2013, ApJ, 778, 10 [NASA ADS] [CrossRef] [Google Scholar]
- Surana, S., Wadadekar, Y., Bait, O., & Bhosale, H. 2020, MNRAS, 493, 4808 [NASA ADS] [CrossRef] [Google Scholar]
- Tagliaferri, R., Longo, G., Andreon, S., et al. 2003, Lect. Notes Comput. Sci., 2859, 226 [NASA ADS] [CrossRef] [Google Scholar]
- Tremonti, C. A., Heckman, T. M., Kauffmann, G., et al. 2004, ApJ, 613, 898 [Google Scholar]
- Turner, S., Kelvin, L. S., Baldry, I. K., et al. 2019, MNRAS, 482, 126 [NASA ADS] [CrossRef] [Google Scholar]
- Vasconcellos, E. C., de Carvalho, R. R., Gal, R. R., et al. 2011, AJ, 141, 189 [Google Scholar]
- Walcher, J., Groves, B., Budavári, T., & Dale, D. 2011, Ap&SS, 331, 1 [NASA ADS] [CrossRef] [Google Scholar]
- Walmsley, M., Smith, L., Lintott, C., et al. 2020, MNRAS, 491, 1554 [Google Scholar]
- Wijesinghe, D. B., Hopkins, A. M., Brough, S., et al. 2012, MNRAS, 423, 3679 [NASA ADS] [CrossRef] [Google Scholar]
- Wright, E. L., Eisenhardt, P. R. M., Mainzer, A. K., et al. 2010, AJ, 140, 1868 [Google Scholar]
- York, D. G., Adelman, J., Anderson, Jr., J. E., et al. 2000, AJ, 120, 1579 [NASA ADS] [CrossRef] [Google Scholar]
- Zeraatgari, F. Z., Hafezianzadeh, F., Zhang, et al. 2024, MNRAS, 527, 4677 [Google Scholar]
- Zhan, H. 2011, Sci. Sin. Phys. Mech. Astron., 41, 1441 [NASA ADS] [CrossRef] [Google Scholar]
All Tables
Comparative evaluation metrics of ML models (CatBoost and WDNN) for the physical parameters SFR, SM, and metallicity.
Comparison of CatBoost model performance on physical parameters with and without observational errors in input features but using the same dataset.
All Figures
![]() |
Fig. 1 Distribution of SFR (left panel), SM (middle panel), and metallicity (right panel). |
In the text |
![]() |
Fig. 2 Performance evaluation of CatBoost for different galaxy property predictions with a varying ‘depth’ based on the mean absolute error for SFR (left panel), SM (middle panel), and metallicity (right panel). The results presented are the average performance across a ten-fold cross-validation. |
In the text |
![]() |
Fig. 3 Schematic structure of the WDNN models utilised in this study. The input features consist of a combination of eight optical and IR magnitudes and colours. The left panel outputs SFR and SM, while the right panel focuses on predicting metallicity. |
In the text |
![]() |
Fig. 4 Feature importance analysis by CatBoost. The left panel demonstrates the significance of features for predicting SFR and SM, while the right panel provides insights into feature importance for predicting metallicity. |
In the text |
![]() |
Fig. 5 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of SFR estimates from the ML model with SFR values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the SFR errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
In the text |
![]() |
Fig. 6 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of SM estimates from the ML model with SM values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the SM errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
In the text |
![]() |
Fig. 7 Results of CatBoost (left panels) and WDNN (right panels) on the test sample (20% of the entire dataset). The upper row subplots present a comparison of metallicity estimates from the ML model with metallicity values obtained from the SDSS MPA-JHU DR8 catalogue. Errors in the results obtained by the ML models about the true values are displayed at the bottom of the subplots. The pink-shaded region indicates a 3σ scatter of the metallicity errors, with any data point falling outside these limits considered an outlier. In the bottom row, box plots illustrate errors across redshift values. In the residual plots, the dashed lines illustrate the best fit through the scatter plots. |
In the text |
![]() |
Fig. 8 Results of CatBoost in predicting SFR and SM for different kinds of galaxies. |
In the text |
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.