1. Introduction
One of the key factors in investigating galaxy formation and evolution is defining the spin of the galaxy (rotation). As outlined by [1], the spin parameter () provides a robust way to quantify the stellar angular momentum in galaxies. This parameter enables a clear distinction between fast and slow rotators, representing different evolutionary paths and formation mechanisms [2,3,4]. Fast rotators, characterized by high values, are thought to have formed through gas-rich minor mergers or secular processes, including the inflow of external gas [5]. In contrast, slow rotators, characterized by low values, are believed to result from dissipationless processes, such as dry minor mergers [6,7]. Furthermore, they note that the distribution of values in different environments can provide insights into the role of galaxy interactions and environmental effects on galaxy evolution [8,9,10,11,12,13].
The methods for measuring galaxy spin have evolved significantly over the years, encompassing both observational and theoretical approaches. A significant advancement came with the introduction of the parameter [14], which reflects the ratio of rotation velocity to velocity dispersion and was introduced as an indicator of observed rotation [15]. In this method, integrated quantities such as and are measured, representing the sky-averaged values of V (velocity) and (velocity dispersion)1 as observed using integral field spectrographs such as SAURON [1]. However, this method has limitations. For instance, galaxies such as NGC 3379, which exhibits global rotation, and NGC 5813, which shows spatially confined non-zero velocities, yield similar values despite having distinct kinematic structures. On the other hand, the advent of integral field spectroscopy enabled spatially resolved kinematics to be established across entire galaxies.This led to the development of the parameter by [1], which provided a more robust measure of specific angular momentum.
The origin of angular momentum captured by the kinematic parameter has been a topic of significant theoretical interest. For instance, ref. [16] proposes an explanation for the origin of angular momentum in dark matter halos. On the theoretical front, numerical simulations have played a crucial role in understanding the origins and evolution of galaxy spin. In particular, ref. [17] conducted hydrodynamic simulations to show that the topology of the merging regions (specifically, the number of intersecting filaments) accurately predicts the spin of both dark matter and gas. They found that halos located at the centers of knots exhibit low spin, whereas those at the centers of filaments exhibit high spin.
Upcoming sky surveys, such as those that will be performed by the Hector instrument [18,19] and the Wide-field Spectroscopic Telescope (WST) [20], are expected to provide even larger databases with billions of galaxies. However, identifying slow/fast rotators within such an enormous number of galaxies cannot be achieved using traditional statistical methods and requires specialized tools. Convolutional neural networks (CNNs) [21], a branch of machine learning models, have gained prominence in cosmology and astronomy. Their power lies in elucidating patterns within complex datasets, and they have proven to be highly effective for diverse scientific tasks. In the field of exoplanet detection, CNN methods have been employed to enhance the process of identification and examination of potential exoplanets from large-scale datasets [22,23]. In radio astronomy, deep learning approaches such as DECORAS [24] have been applied to differentiate and characterize radio-astronomical sources. They have also been helpful in processing data on active galactic nuclei (AGNs) [25], for example, detecting and classifying AGN host galaxies [26]. They have also been valuable in strong lensing searches [27,28,29], improving efficiency by reducing the time required while enhancing accuracy. Furthermore, CNNs have proven to be valuable in cosmology, showing potential to address computational limitations faced by conventional statistical methods in research on dark energy [30,31,32], dark matter [33,34,35], Cosmic Microwave Background (CMB) mapping [36,37,38], and Gravitational Waves (GWs) [25,39,40]. These applications showcase the remarkable versatility and capability of CNNs in deepening our understanding of the universe. Refs. [41,42,43] provide an extensive overview of CNN applications to address a wide range of astronomical and cosmological challenges.
In this study, we leverage CNNs to classify galaxies as slow or fast rotators based on their stellar kinematic maps. We adopt a supervised learning approach, utilizing labeled data from the SAMI catalog as our training, test, and validation datasets.2 Through this process, we seek to find the best CNN architecture, and we achieve an accuracy of approximately 91% on the test dataset. Subsequently, we apply our trained model to classify previously unknown rotator galaxies whose rotation could not be classified by traditional statistical tools as fast or slow rotation. We utilized interpretability techniques, including Integrated Gradients (IGs), to uncover the key kinematic features that guided the CNN’s classification decisions. Ultimately, this work highlights the effectiveness of CNNs in distinguishing between slow and fast rotators, demonstrating their potential to advance our understanding of galaxy dynamics and informing future observational strategies.
The structure of this paper is organized as follows. In Section 2, we describe the source, preprocessing steps, and preparation of the SAMI survey [44,45] data, which are used for the CNN-based classification of slow and fast galaxy rotators. In this section, we also provide a detailed explanation of the network architecture, data preprocessing, and evaluation methods. Our results and their physical interpretations are discussed in Section 3. In this section, we also present a series of studies aimed at elucidating how our CNN model makes decisions when classifying images. Finally, we discuss and summarize the main results of this work in Section 4.
2. Dataset and Machine Learning Methods
The data source used in this study is the SAMI survey. The SAMI instrument [44] was installed on the 3.9-meter Anglo-Australian Telescope (AAT) and connected to the AAOmega spectrograph [46,47,48,49,50,51]. This setup provides a median resolution of in the range of and in the range of [52]. We concentrate on a sample that includes stellar kinematics, selecting only the Brightest Group Galaxies (BGGs) with stellar masses of , ensuring the inclusion of massive galaxies with high signal-to-noise ratios (S/N > 5). In our study, we processed a dataset comprising stellar kinematics data on 2444 galaxies from the SAMI survey. We systematically iterated the process through the directory containing the galaxies’ FITS (Flexible Image Transport System) files, extracting relevant information from each file. Specifically, we retrieved the primary data array from each FITS file and associated it with a unique identifier derived from its filename. Then, we stored these data arrays and their corresponding identifiers in separate files.
For our data processing pipeline, we applied several crucial steps to ensure data quality and relevance for our analysis of galaxy stellar kinematics. First, we addressed the issue of duplicated labels in our dataset and removed any duplicate entries from our label array, ensuring that each galaxy in our sample was represented uniquely. Thereafter, we handled missing values in our stellar kinematics data by replacing NaN values with zeros, a common practice in astronomical data processing when dealing with regions of low signal-to-noise ratios. We categorized galaxies based on their rotational properties. Based on the spin–ellipticity relations from [53], we categorized each galaxy as either a fast rotator (designated as 1) or a slow rotator (designated as 0). Finally, to standardize our spatial analysis, we extracted a central region of each galaxy’s kinematic map, creating a consistent spatial scale across our sample. By selecting a square region of pixels centered on each galaxy, we ensured that our analysis focused on the most relevant and well-measured parts of each galaxy while maintaining uniform spatial coverage across our sample. These data processing steps were crucial in preparing a clean, consistent, and well-characterized dataset for making our training, test and validation datasets. These datasets can used in our CNN binary classification problem. Then, we applied a CNN to classify stellar kinematic maps that were previously labeled as fast or slow rotators.
2.1. Convolutional Neural Network Architecture
The use of CNNs is a highly effective and widely used technique for object recognition and classification in the field of Machine Learning (ML) [54]. These neural networks employ a hierarchical architecture consisting of multiple layers, including convolutional and pooling operations. Through this architecture, a CNN can automatically learn and extract meaningful features from input images, enabling it to process complex visual data with remarkable accuracy [55]. In our study, we applied a CNN to two-dimensional stellar kinematic maps to distinguish between fast and slow rotators.
We employed a systematic approach to design an optimal CNN architecture for classifying galaxies based on their stellar kinematic properties. We utilized Keras Tuner [56], an automated hyperparameter optimization framework, to explore a wide range of model configurations efficiently. We varied the number of convolutional layers (between 1 and 4), the number of filters in each layer (from 32 to 256 in steps of 32), and the learning rate (chosen from , , and ). This approach enabled us to systematically evaluate different model complexities and training parameters. The resulting optimal architecture, as determined by Keras Tuner, is shown in Table 1.
In the CNN architecture, our goal was to minimize the binary cross-entropy loss. To achieve this, we employed the Adam optimizer [57] with a learning rate of 0.0002 and a beta parameter () of 0.9. Additionally, we implemented two checkpoints in the model. The first one was ModelCheckpoint, which saves the best model during training based on the validation accuracy. The second one was ReduceLROnPlateau, which reduces the learning rate when there is no progress in training. Moreover, we incorporated the EarlyStopping callback, which stops the training process if the validation accuracy does not improve. To monitor the training progress, we recorded the loss function and accuracy computed on the training data after each epoch. We also calculated the validation loss and accuracy on a separate validation dataset. These metrics provided insights into the performance of the model during training and helped track its progress.
2.2. Data Preprocessing
We enhanced our dataset using a custom data augmentation function. This function takes the original data and labels as input and applies various transformations to generate additional training samples, thereby increasing the diversity of the training set. Specifically, we utilized the ImageDataGenerator from Keras, configured with parameters such as a rotation range of 360 degrees, width and height shifts of up to 4 pixels, a zoom range of 1.0 to 1.05, and horizontal flipping, with the fill mode set to nearest to maintain the integrity of the images. To balance the numbers of fast and slow rotator samples in the dataset, the function performed seven times more augmentations on the slow rotator samples. These transformed images were added to the dataset along with their corresponding labels, ensuring an equal representation of fast and slow rotators.
After preprocessing, the dataset consisted of 1112 samples, with 980 classified as fast rotators and 132 as slow rotators. To address the class imbalance, we applied data augmentation, specifically aimed at balancing the number of slow and fast rotators. As a result, the final dataset comprised 19,332 samples, with 9603 fast rotators and 9729 slow rotators, ensuring a more balanced distribution for training the model. Then we divided it into three subsets: 70% for training, 15% for testing, and the remainder for validation. This split ensured that the model was trained on a substantial portion of the data while also being evaluated and validated on separate, unseen subsets to gauge its performance and generalization capabilities. This augmentation and splitting strategy was crucial for improving the robustness and accuracy of our CNN in classifying stellar kinematic maps as belonging to fast or slow rotators.
2.3. Loss Function
Loss functions were used to evaluate our network’s accuracy by measuring the inequality between predicted and true class labels. For our binary classification problem of distinguishing fast from slow rotators, we employed the Binary Cross-Entropy (BCE) [58] loss function. Optimization was carried out using the Adam optimizer with a learning rate of 0.001, and the BCE loss function is formulated as follows:
In this context, corresponds to the actual class label of the ith sample within our dataset of N training samples, and is the probability predicted by the model that the ith sample is a fast or slow rotator.2.4. Evaluation Criteria
To evaluate and compare the effectiveness of the CNN on the test dataset, it is crucial to establish appropriate evaluation criteria. In the context of stellar kinematic maps, which is treated as a classification problem distinguishing between fast and slow rotator samples, we can utilize the following confusion matrix for evaluation purposes:
In this context, True Positive (TP) indicates a correctly identified fast rotator, True Negative (TN) corresponds to an accurately recognized slow rotator, False Positive (FP) denotes misclassification of a slow rotator source as fast, and False Negative (FN) refers to a fast rotator missed by the algorithm and classified as slow.
These terms allow us to define various evaluation metrics. Accuracy measures the fraction of correctly identified samples (TP and TN) out of the total number of samples. Precision quantifies the ratio of true positives to the sum of true positives and false positives, showing the reliability of positive predictions [59]. These metrics are defined as follows:
(1)
The Receiver Operating Characteristic (ROC) curve [60] is another useful metric, depicting the trade-off between TPs and FPs for the CNN model. Each point on the ROC curve corresponds to a different threshold for categorizing samples as slow or fast rotators based on their predicted probabilities. By analyzing the ROC curve, we can determine how well the model distinguishes between fast and slow rotator samples. A model with superior performance will have a curve that closely approaches the top left corner of the plot, signifying a higher TP rate and a lower FP rate for various thresholds.3. Results
In this section, we present the results that show the performance of our CNN-based model in classifying fast and slow rotators using the stellar kinematic maps from the SAMI survey. We first investigate the model’s classification performance through the confusion matrix and the ROC curve and analyze the distribution of the model’s predictions. Afterward, we apply the CNN model to unknown rotational samples and analyze the key features of the stellar maps that influence the model’s decision-making process. To achieve this, we utilize an explainable CNN approach, including Integrated Gradients (IGs).
3.1. Training CNN on Galaxies with Known Fast/Slow Rotation
Using SAMI observations, the stellar kinematic maps of galaxies can be estimated and the parameter is used as a proxy for the spin parameter. Following [1] it is calculated using:
(2)
where the subscript i refers to the ith spaxel within the ellipse, is the flux of the ith spaxel, is the stellar velocity in km s−1, and is the velocity dispersion in km s−1. We trained the CNN on training data including stellar kinematic maps of 7697 fast and 7835 slow rotators, defined by the parameter .In Figure 1, we show the confusion matrix for the CNN model, which demonstrates the strong performance of the model, especially in terms of precision, suggesting that it effectively minimizes false positives for the slow rotator category. The overall accuracy further reinforces the model’s reliability in making predictions between the two classes. Since the model output represents the probability of an object being a fast rotator, ranging between 0 and 1, setting a threshold is necessary to classify objects into two classes: fast and slow rotators. Figure 2 illustrates the ROC plot, which has a steep ascent, quickly reaching high TP rates even at low FP rates, a pattern that is indicative of a strong classifier. The Area Under the ROC Curve (AUC) is reported as 0.95, which is very close to the ideal value of 1.0. This high AUC score suggests that the model has excellent discriminative ability between the slow and fast rotator categories. Figure 3 reveals the distributions of both the CNN predictions (blue) and the real data (pink). There are two prominent peaks at the extreme ends (0 and 1) of both distributions, indicating that the model is making confident predictions for a large number of samples, classifying them as either very likely slow rotators (close to 0) or very likely fast rotators (close to 1). The CNN’s predictions closely align with the real data distribution, especially at the extremes. This suggests that the model is performing well in identifying clear-cut cases. However, there are some discrepancies in the middle range (around 0.2 to 0.8), where the CNN shows more varied predictions compared to the real data. This could indicate that the model is less certain about borderline cases and produces a wider range of probability estimates for these samples.
Figure 4 displays different visualizations of false positive samples from the CNN model. Each visualization represents a different instance where the model incorrectly classified a slow rotator sample as a fast one. Figure 5 displays visualizations of false negative samples from the CNN model. Each visualization represents a different instance where the model incorrectly classified a fast rotator sample as a slow one. These false positive and false negative cases are crucial for understanding where and why the model is making mistakes. They could represent edge cases or challenging inputs that share some characteristics with positive samples, leading to misclassification. Analyzing these false positives can provide insights into the model’s decision-making process and highlight areas where it might be overfitting or misinterpreting certain features. This information is valuable for refining the model, adjusting its architecture, and enhancing the training data to improve overall performance and reduce such misclassifications in future iterations.
3.2. Testing the CNN on Unknown Rotators
Figure 6 shows the CNN-predicted probability distribution for unknown fast and slow rotator samples, while Figure 7 presents the stellar maps of some of these samples. Traditional techniques that rely on the spin parameter derived from stellar kinematics often face significant challenges, including the misclassification of galaxies with complex rotation patterns and the inability to effectively analyze low-surface-brightness galaxies [61]. These limitations stem from the conventional methods’ struggles to capture the intricate details of stellar motion across diverse galaxy environments. For example, galaxies with non-axisymmetric features or those affected by interactions with neighboring galaxies can yield misleading values. Moreover, low-surface-brightness galaxies, which often exhibit considerable structural complexity, may fall below the detection thresholds of standard techniques, resulting in gaps in our understanding of their dynamics.
Among the 1112 samples in the SAMI catalogue, a significant portion (854 samples) were not classified as either slow or fast rotators due to these limitations. The CNN’s ability to learn from extensive datasets enables it to discern and classify intricate rotational signatures, thereby enhancing the accuracy of galaxy classification. This methodology not only improves our ability to categorize fast and slow rotators but also allows us to investigate previously ambiguous cases involving complex rotations or low-surface-brightness galaxies, ultimately advancing our comprehension of galaxy formation and evolution.
3.3. Interpretability of the Model’s Classifications
This subsection outlines a series of studies aimed at elucidating how our CNN model arrives at its decisions when classifying images. Two primary techniques are used to gain insight into the decision-making processes of the CNN model:
(i). Measuring the clustering of velocity variations in the stellar kinematic maps using the Global Moran’s I (GMI) tool [62,63];
(ii). Utilizing Integrated Gradients (IGs) [64] to emphasize the areas of input images that significantly correspond to the fast or slow classes.
We analyzed the histograms of GMI values for the fast and slow rotators, emphasizing the significance of clustering among high- and low-velocity stars in defining these categories. Afterwards, we employed IGs to identify key regions within the velocity stellar maps that influence the decision-making process of our CNN model. Finally, we investigated the model’s decisions for true positives and true negatives, exploring the relationship between clustering in velocities of stars and the IG values outcomes.
3.3.1. Clustering of High- and Low-Velocity Stars
Visual inspection of the velocity variations in the stellar kinematic maps revealed a more pronounced clustering of high- and low-velocity stars in fast rotators than in slow rotators. Accordingly, we utilized the GMI tool to assess the spatial autocorrelation of stellar maps by considering both the locations and values of features simultaneously. Specifically, these observed patterns indicate a clustering of high- and low-velocity stars within the stellar maps. GMI index values measure the spatial autocorrelation, which quantifies how similar or dissimilar data values are based on their spatial proximity. It is often used in spatial statistics to determine whether there is clustering, dispersion, or randomness in a data sample. The formula for the GMI index is as follows:
(3)
where n is the number of observations, is the value of the variable of interest at location i, is the mean of the variable, and is the spatial weight between locations i and j. The interpretation of the GMI index value can be categorized as follows:(i). A positive GMI index (close to +1) indicates clustering or positive spatial autocorrelation, meaning similar values (either high or low) tend to be close to each other;
(ii). A negative GMI index (close to −1) indicates dispersion or negative spatial autocorrelation, suggesting that high and low values are intermixed and tend to avoid clustering;
(iii). A zero or near-zero GMI index suggests a random spatial pattern with no clear clustering or dispersion.
Figure 8 shows the histograms of GMI values for fast and slow rotators. The distribution shows that there are significantly more slow rotators in the lower ranges (0.0 to 0.4) of Moran’s I. The fast rotators feature more prominently in the middle to higher ranges (approximately 0.4 to 1.0), suggesting different distribution patterns between the two categories. In Figure 9, we show the spatial autocorrelation in star velocity maps for two fast rotators from the SAMI dataset, which is predicted correctly by our CNN model. The central scatter plot presents the Moran scatter plot for each fast rotator, where the spatial lag of standardized velocity is plotted against the standardized star velocity. High Moran’s I values (0.910 and 0.930) indicate strong spatial autocorrelation in all cases, suggesting coherent kinematic structures within these fast rotators. The top row shows spatial velocity maps, with red and blue regions indicating higher and lower velocities, respectively. In contrast, Figure 10 shows the previous analysis for the two slow rotators. Lower Moran’s I values (0.385, and 0.180) suggest weaker spatial autocorrelation, indicating less coherent kinematic structures compared to the fast rotators.
The examination of stellar kinematics and the Global Moran’s I (GMI) index reveals substantial differences between fast and slow rotators in terms of velocity distribution. Fast rotators exhibit more distinct clustering of both high and low velocities, with GMI values between 0.4 and 1.0, indicating a higher degree of spatial autocorrelation and more uniform kinematic structures. Conversely, slow rotators display lower GMI values (0.0 to 0.4), pointing to weaker spatial autocorrelation and less coherent velocity distributions. This suggests that fast rotators have more cohesive velocity arrangements than slow rotators, which exhibit more scattered and irregular distributions.
3.3.2. Integrated Gradients
The Integrated Gradient (IG) method [64] is a systematic method used to evaluate the contribution of individual pixels or regions in an input image to the model’s final classification decision. To thoroughly understand the IG method, it is crucial to explore the concept of alpha, a key parameter in this approach. Alpha is a real number ranging from 0 to 1. It establishes a path along which the gradients of the model’s predicted probabilities are evaluated. Specifically, dictates the proportional contribution of information from a baseline image relative to the target image being analyzed. This proportional contribution is calculated as an integral along the defined path. Recently, IGs have had numerous applications in astronomy, including strong lensing searches [65] and weak lensing analysis [66].
IGs are applied to the slow and fast rotators in Figure 9 and Figure 10, respectively. In Figure 9, in the right row of each sample, we show the IG value for each pixel of these samples. It should be noted that pixels with IG values greater than zero have positive effect on our model’s decision and vice versa. The majority of these IG images highlight the pixels in the coherent kinematic structures in the image. Thus, according to this analysis we can say one of the CNN model features used in distinguishing fast and slow rotators is the coherent kinematic structures in the fast rotators, while in Figure 10, showing slow rotators’ IG images, the majority of pixels are concentrated at a weaker spatial autocorrelation, indicating less coherent kinematic structures.
4. Discussion and Conclusions
In this work, we have demonstrated the success of convolutional neural networks (CNNs) in effectively classifying galaxies into fast and slow rotators based on their stellar kinematic maps. Using data from the SAMI survey, we developed and fine-tuned a CNN architecture that delivered an accuracy of 91% and a precision of 95%, proving its reliability in distinguishing between the two categories. These results illustrate the potential of CNNs to overcome challenges in galaxy classification, particularly when traditional methods relying on spin parameters encounter uncertainties.
The analysis of stellar kinematics using the Global Moran’s I (GMI) index reveals distinct differences between the velocity distributions of fast and slow rotators. In particular, the analysis suggests that fast rotators have more organized and structured velocity fields, whereas slow rotators exhibit more scattered and irregular distributions, highlighting fundamental differences in the kinematic properties of slow and fast rotators. The interpretability of CNN classifications is an important component of this study. By applying Integrated Gradients, we uncovered the primary kinematic feature that significantly affected the model’s outputs. Our findings indicate that the CNN focuses on coherent kinematic structures within stellar velocity maps to classify galaxies, confirming its ability to detect meaningful physical patterns. Using this interpretable CNN model, we can enhance confidence in the model’s predictions.
The scalability and adaptability of CNNs are essential given the emergence of current and future large-scale integral field spectroscopy (IFS) surveys, such as the Hector instrument [18,19] (which is designed to characterize 15,000 galaxies) and the Wide-field Spectroscopic Telescope (WST) [20], which generate datasets comprising billions of galaxies. Under this condition, traditional statistical approaches will face significant challenges in handling such extensive data effectively. Our findings highlight the vital role of CNNs for processing and analyzing stellar kinematic patterns in these large datasets, establishing their importance for the next generation of astronomical exploration.
Although the findings are promising, the study highlights some limitations. The model showed uncertainty in borderline cases, as shown in the probability distribution of predictions. Addressing this will require larger and more diverse training datasets. Future research could aim to include expanding the approach to account for advanced kinematic features, such as higher-order velocity moments, or leveraging transfer learning with models pre-trained on similar datasets. Furthermore, employing more explainable CNN models allows us to identify the most significant features in the decision-making process. This understanding can be leveraged to enhance the model’s accuracy by incorporating attention layers [67], which focus on the most relevant aspects of the data. We intend to broaden our analysis by investigating CNNs alongside traditional machine learning classifiers, including random forests [68], support vector machines [69], and Naive Bayes [70], with a focus on both accuracy and model interpretability.
The first two authors are considered the main contributors to this paper and are listed as co-first authors, alphabetized by first name. Conceptualization: A.C. and F.F.H.; Methodology: A.C. and F.F.H.; Software: M.R.; Validation: M.R., B.F. and F.J.V.; Formal Analysis: A.C. and F.F.H.; Investigation: F.F.H.; Resources: M.R.; Data Curation: M.R.; Writing—Original Draft Preparation: M.R.; Writing—Review and Editing: M.R.; Visualization: A.C.; Supervision: M.R., F.J.V. and B.F.; Project Administration: A.C. and F.F.H.; Funding Acquisition: F.F.H. All authors have read and agreed to the published version of the manuscript.
The code used in this study is publicly available on GitHub in the following repository:
F.F. and M.R. would like to express their heartfelt appreciation to EuroSpaceHub and LUNEX EuroMoonMars Earth Space Innovation for their generous funding and unwavering support. A.C.H. was supported by MUR PRIN2022 project 20222JBEKN, titled “LaScaLa”, funded by the European Union—NextGenerationEU.
The authors declare no conflicts of interest.
Footnotes
1. The angle brackets indicate a sky average weighted by surface brightness.
2. We obtained results similar to those from the MaNGA dataset; however, we prefer to focus on the SAMI survey for this study, as it offers the advantage of higher signal-to-noise ratios in the stellar kinematics of galaxies compared to MaNGA. While we did apply our method to a small sample of MaNGA data with known
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1. Confusion matrix illustrating the model’s performance, with an accuracy of 91% and a precision of 95%, highlighting its effectiveness in classifying slow and fast rotators.
Figure 2. The ROC plot demonstrates how the CNN model perform in distinguishing between fast and slow rotator samples across various thresholds, balancing true positive (TP) and false positive (FP) rates. The plot uses a logarithmic scale for the x-axis (FP rate), allowing a more detailed view of the model’s performance at low FP rates.
Figure 3. This distribution plot compares the predictions made by the CNN model to the real test data. The x-axis represents the prediction probability from 0 to 1, while the y-axis shows the number of samples on a logarithmic scale.
Figure 4. These images presents distinct visualizations of samples for which the CNN model gave false positives.Each visualization illustrates a different instance where the model mistakenly identified a slow rotator sample as a fast rotator.
Figure 5. These images presents distinct visualizations of samples for which the CNN gave false negatives.Each visualization illustrates a different instance where the model mistakenly identified a fast rotator sample as a slow rotator.
Figure 6. Histogram of CNN prediction probabilities (P) for unknown samples. The distribution is marked by two threshold regions: fast rotators ([Forumla omitted. See PDF.], red dashed line) and slow samples ([Forumla omitted. See PDF.], blue dashed line). The total numbers of fast ([Forumla omitted. See PDF.]) and slow ([Forumla omitted. See PDF.]) samples are 678 and 39, respectively.
Figure 7. These images showcase various visualizations of the unknown rotator samples for which it was significantly challenging to derive the [Forumla omitted. See PDF.] spin parameter. Our CNN model was applied to these images, and the probability predicted by the CNN is reported on the top of each image.
Figure 8. Distributions of Moran’s I for slow and fast rotators, illustrating differences in spatial autocorrelation. Slow rotators (blue) exhibit a narrower distribution centered around lower values, while fast rotators (red) display a broader distribution at higher values.
Figure 9. Spatial autocorrelation of star velocities for two fast rotators from the SAMI dataset as predicted by our CNN model. The central scatter plot shows the Moran scatter plot for each fast rotator, where the spatial lag of standardized velocity is plotted against the standardized star velocity. The top row presents spatial velocity maps, with red and blue regions representing higher and lower velocities, respectively. The right row displays the Integrated Gradient (IG) value for each pixel, where positive IG values contribute positively to the model’s decision.
Figure 10. Spatial autocorrelation of star velocities for two slow rotators from the SAMI dataset. The central scatter plot shows the Moran scatter plot for each slow rotator, where the spatial lag of standardized velocity is plotted against the standardized star velocity. Lower Moran’s I values (0.385 and 0.180) suggest weaker spatial autocorrelation, indicating less coherent kinematic structures compared to the fast rotators analyzed in Figure 9. The bottom row displays the Integrated Gradient (IG) value for each pixel, highlighting that the majority of integrated gradient values are concentrated in regions with weaker spatial autocorrelation, further indicating less coherent kinematic structures in the slow rotators.
Summary of the convolutional neural network (CNN) architecture, including layer types, output shapes, and parameter counts.
Layer Type | Output Shape | Parameters |
---|---|---|
Conv2D | (None, 38, 38, 64) | 640 |
MaxPooling2D | (None, 19, 19, 64) | 0 |
BatchNormalization | (None, 19, 19, 64) | 256 |
Conv2D | (None, 17, 17, 128) | 73,856 |
MaxPooling2D | (None, 8, 8, 128) | 0 |
BatchNormalization | (None, 8, 8, 128) | 512 |
Conv2D | (None, 6, 6, 256) | 295,168 |
MaxPooling2D | (None, 3, 3, 256) | 0 |
BatchNormalization | (None, 3, 3, 256) | 1024 |
Flatten | (None, 2304) | 0 |
Dense | (None, 96) | 221,280 |
Dropout | (None, 96) | 0 |
Dense | (None, 32) | 3104 |
Dropout | (None, 32) | 0 |
Dense | (None, 1) | 33 |
References
1. Emsellem, E.; Cappellari, M.; Krajnović, D.; Van De Ven, G.; Bacon, R.; Bureau, M.; Davies, R.L.; De Zeeuw, P.; Falcón-Barroso, J.; Kuntschner, H. et al. The SAURON project—IX. A kinematic classification for early-type galaxies. Mon. Not. R. Astron. Soc.; 2007; 379, pp. 401-417. [DOI: https://dx.doi.org/10.1111/j.1365-2966.2007.11752.x]
2. Raouf, M.; Smith, R.; Khosroshahi, H.G.; Sande, J.v.d.; Bryant, J.J.; Cortese, L.; Brough, S.; Croom, S.M.; Hwang, H.S.; Driver, S. et al. The SAMI Galaxy Survey: Kinematics of Stars and Gas in Brightest Group Galaxies—The Role of Group Dynamics. Astrophys. J.; 2021; 908, 123. [DOI: https://dx.doi.org/10.3847/1538-4357/abd47d]
3. Raouf, M.; Viti, S.; García-Burillo, S.; Richings, A.J.; Schaye, J.; Bemis, A.; Nobels, F.S.J.; Guainazzi, M.; Huang, K.Y.; Schaller, M. et al. Hydrodynamic simulations of the disc of gas around supermassive black holes (HDGAS)—I. Molecular gas dynamics. Mon. Not. R. Astron. Soc.; 2023; 524, pp. 786-800. [DOI: https://dx.doi.org/10.1093/mnras/stad1957]
4. Raouf, M.; Purabbas, M.H.; Fazel Hesar, F. Impact of Active Galactic Nuclei Feedback on the Dynamics of Gas: A Review across Diverse Environments. Galaxies; 2024; 12, 16. [DOI: https://dx.doi.org/10.3390/galaxies12020016]
5. Cappellari, M.; Emsellem, E.; Bacon, R.; Bureau, M.; Davies, R.L.; De Zeeuw, P.; Falcón-Barroso, J.; Krajnović, D.; Kuntschner, H.; McDermid, R.M. et al. The SAURON project—X. The orbital anisotropy of elliptical and lenticular galaxies: Revisiting the (V/σ, ε) diagram with integral-field stellar kinematics. Mon. Not. R. Astron. Soc.; 2007; 379, pp. 418-444. [DOI: https://dx.doi.org/10.1111/j.1365-2966.2007.11963.x]
6. Hernquist, L. Structure of merger remnants. II-Progenitors with rotating bulges. Astrophys. J.; 1993; 409, pp. 548-562. [DOI: https://dx.doi.org/10.1086/172686]
7. Burkert, A.; Naab, T.; Johansson, P.H.; Jesseit, R. SAURON’s Challenge for the Major Merger Scenario of Elliptical Galaxy Formation. Astrophys. J.; 2008; 685, 897. [DOI: https://dx.doi.org/10.1086/591632]
8. Blanton, M.R.; Moustakas, J. Physical properties and environments of nearby galaxies. Annu. Rev. Astron. Astrophys.; 2009; 47, pp. 159-210. [DOI: https://dx.doi.org/10.1146/annurev-astro-082708-101734]
9. Naab, T.; Oser, L.; Emsellem, E.; Cappellari, M.; Krajnović, D.; McDermid, R.M.; Alatalo, K.; Bayet, E.; Blitz, L.; Bois, M. et al. The ATLAS3D project—XXV. Two-dimensional kinematic analysis of simulated galaxies and the cosmological origin of fast and slow rotators. Mon. Not. R. Astron. Soc.; 2014; 444, pp. 3357-3387. [DOI: https://dx.doi.org/10.1093/mnras/stt1919]
10. Raouf, M.; Khosroshahi, H.G.; Ponman, T.J.; Dariush, A.A.; Molaeinezhad, A.; Tavasoli, S. Ultimate age-dating method for galaxy groups; clues from the Millennium Simulations. Mon. Not. R. Astron. Soc.; 2014; 442, pp. 1578-1585. [DOI: https://dx.doi.org/10.1093/mnras/stu963]
11. Raouf, M.; Khosroshahi, H.G.; Dariush, A. Evolution of Galaxy Groups in the Illustris Simulation. Astrophys. J.; 2016; 824, 140. [DOI: https://dx.doi.org/10.3847/0004-637X/824/2/140]
12. Raouf, M.; Khosroshahi, H.G.; Mamon, G.A.; Croton, D.J.; Hashemizadeh, A.; Dariush, A.A. Merger History of Central Galaxies in Semi-analytic Models of Galaxy Formation. Astrophys. J.; 2018; 863, 40. [DOI: https://dx.doi.org/10.3847/1538-4357/aace57]
13. Raouf, M.; Smith, R.; Khosroshahi, H.G.; Dariush, A.A.; Driver, S.; Ko, J.; Hwang, H.S. The Impact of the Dynamical State of Galaxy Groups on the Stellar Populations of Central Galaxies. Astrophys. J.; 2019; 887, 264. [DOI: https://dx.doi.org/10.3847/1538-4357/ab5581]
14. Illingworth, G. Rotation in 13 elliptical galaxies. Astrophys. J.; 1977; 218, pp. L43-L47. [DOI: https://dx.doi.org/10.1086/182572]
15. Binney, J. Rotation and anisotropy of galaxies revisited. Mon. Not. R. Astron. Soc.; 2005; 363, pp. 937-942. [DOI: https://dx.doi.org/10.1111/j.1365-2966.2005.09495.x]
16. Vitvitska, M.; Klypin, A.A.; Kravtsov, A.V.; Wechsler, R.H.; Primack, J.R.; Bullock, J.S. The origin of angular momentum in dark matter halos. Astrophys. J.; 2002; 581, 799. [DOI: https://dx.doi.org/10.1086/344361]
17. Prieto, J.; Jimenez, R.; Haiman, Z.; González, R.E. The origin of spin in galaxies: Clues from simulations of atomic cooling haloes. Mon. Not. R. Astron. Soc.; 2015; 452, pp. 784-802. [DOI: https://dx.doi.org/10.1093/mnras/stv1234]
18. Bryant, J.J.; Bland-Hawthorn, J.; Lawrence, J.; Norris, B.; Min, S.S.; Brown, R.; Wang, A.; Bhatia, G.S.; Saunders, W.; Content, R. et al. Hector: A new multi-object integral field spectrograph instrument for the Anglo-Australian Telescope. Proceedings of the Ground-Based and Airborne Instrumentation for Astronomy VIII; Online Conference, 14–18 December 2020; Volume 11447, pp. 201-207.
19. Bryant, J.J.; Oh, S.; Gunawardhana, M.; Quattropani, G.; Bhatia, G.S.; Bland-Hawthorn, J.; Broderick, D.; Brown, R.; Content, R.; Croom, S. et al. Hector: Performance of the new integral field spectrograph instrument for the Anglo-Australian Telescope. Proceedings of the Ground-Based and Airborne Instrumentation for Astronomy X; Yokohama, Japan, 16–21 June 2024; Volume 13096, pp. 88-98.
20. Mainieri, V.; Anderson, R.I.; Brinchmann, J.; Cimatti, A.; Ellis, R.S.; Hill, V.; Kneib, J.P.; McLeod, A.F.; Opitom, C.; Roth, M.M. et al. The Wide-field Spectroscopic Telescope (WST) Science White Paper. arXiv; 2024; [DOI: https://dx.doi.org/10.48550/arXiv.2403.05398] arXiv: 2403.05398
21. Koushik, J. Understanding Convolutional Neural Networks. arXiv; 2016; arXiv: 1605.09081
22. Valizadegan, H.; Martinho, M.J.; Wilkens, L.S.; Jenkins, J.M.; Smith, J.C.; Caldwell, D.A.; Twicken, J.D.; Gerum, P.C.; Walia, N.; Hausknecht, K. et al. ExoMiner: A highly accurate and explainable deep learning classifier that validates 301 new exoplanets. Astrophys. J.; 2022; 926, 120. [DOI: https://dx.doi.org/10.3847/1538-4357/ac4399]
23. Cuéllar, S.; Granados, P.; Fabregas, E.; Curé, M.; Vargas, H.; Dormido-Canto, S.; Farias, G. Deep learning exoplanets detection by combining real and synthetic data. PLoS ONE; 2022; 17, e0268199. [DOI: https://dx.doi.org/10.1371/journal.pone.0268199]
24. Rezaei, S.; McKean, J.P.; Biehl, M.; Javadpour, A. DECORAS: Detection and characterization of radio-astronomical sources using deep learning. Mon. Not. R. Astron. Soc.; 2022; 510, pp. 5891-5907. [DOI: https://dx.doi.org/10.1093/mnras/stab3519]
25. Cuoco, E.; Powell, J.; Cavaglià, M.; Ackley, K.; Bejger, M.; Chatterjee, C.; Coughlin, M.; Coughlin, S.; Easter, P.; Essick, R. et al. Enhancing gravitational-wave science with machine learning. Mach. Learn. Sci. Technol.; 2020; 2, 011002. [DOI: https://dx.doi.org/10.1088/2632-2153/abb93a]
26. Chang, Y.Y.; Hsieh, B.C.; Wang, W.H.; Lin, Y.T.; Lim, C.F.; Toba, Y.; Zhong, Y.; Chang, S.Y. Identifying AGN host galaxies by machine learning with HSC+ WISE. Astrophys. J.; 2021; 920, 68. [DOI: https://dx.doi.org/10.3847/1538-4357/ac167c]
27. He, Z.; Er, X.; Long, Q.; Liu, D.; Liu, X.; Li, Z.; Liu, Y.; Deng, W.; Fan, Z. Deep learning for strong lensing search: Tests of the convolutional neural networks and new candidates from KiDS DR3. Mon. Not. R. Astron. Soc.; 2020; 497, pp. 556-571. [DOI: https://dx.doi.org/10.1093/mnras/staa1917]
28. Cheng, T.Y.; Li, N.; Conselice, C.J.; Aragón-Salamanca, A.; Dye, S.; Metcalf, R.B. Identifying strong lenses with unsupervised machine learning using convolutional autoencoder. Mon. Not. R. Astron. Soc.; 2020; 494, pp. 3750-3765. [DOI: https://dx.doi.org/10.1093/mnras/staa1015]
29. Rezaei, S.; Chegeni, A.; Nagam, B.C.; McKean, J.P.; Baratchi, M.; Kuijken, K.; Koopmans, L.V.E. Reducing false positives in strong lens detection through effective augmentation and ensemble learning. Mon. Not. R. Astron. Soc.; 2025; staf327. [DOI: https://dx.doi.org/10.1093/mnras/staf327]
30. Chegeni, A.; Hassani, F.; Sadr, A.V.; Khosravi, N.; Kunz, M. Clusternets: A deep learning approach to probe clustering dark energy. arXiv; 2023; arXiv: 2308.03517[DOI: https://dx.doi.org/10.1093/mnras/stae1075]
31. Escamilla-Rivera, C.; Quintero, M.A.C.; Capozziello, S. A deep learning approach to cosmological dark energy models. J. Cosmol. Astropart. Phys.; 2020; 2020, 008. [DOI: https://dx.doi.org/10.1088/1475-7516/2020/03/008]
32. Goh, L.; Ocampo, I.; Nesseris, S.; Pettorino, V. Distinguishing coupled dark energy models with neural networks. Astron. Astrophys.; 2024; 692, A101. [DOI: https://dx.doi.org/10.1051/0004-6361/202451099]
33. Khosa, C.K.; Mars, L.; Richards, J.; Sanz, V. Convolutional neural networks for direct detection of dark matter. J. Phys. Nucl. Part. Phys.; 2020; 47, 095201. [DOI: https://dx.doi.org/10.1088/1361-6471/ab8e94]
34. Lucie-Smith, L.; Peiris, H.V.; Pontzen, A. An interpretable machine-learning framework for dark matter halo formation. Mon. Not. R. Astron. Soc.; 2019; 490, pp. 331-342. [DOI: https://dx.doi.org/10.1093/mnras/stz2599]
35. Herrero-Garcia, J.; Patrick, R.; Scaffidi, A. A semi-supervised approach to dark matter searches in direct detection data with machine learning. J. Cosmol. Astropart. Phys.; 2022; 2022, 039. [DOI: https://dx.doi.org/10.1088/1475-7516/2022/02/039]
36. Sadr, A.V.; Farsian, F. Filling in Cosmic Microwave Background map missing regions via Generative Adversarial Networks. J. Cosmol. Astropart. Phys.; 2021; 2021, 012. [DOI: https://dx.doi.org/10.1088/1475-7516/2021/03/012]
37. Ni, S.; Li, Y.; Zhang, X. CMB delensing with deep learning. arXiv; 2023; arXiv: 2310.07358
38. Mishra, A.; Reddy, P.; Nigam, R. Cmb-gan: Fast simulations of cosmic microwave background anisotropy maps using deep learning. arXiv; 2019; arXiv: 1908.04682
39. Razzano, M.; Cuoco, E. Image-based deep learning for classification of noise transients in gravitational wave detectors. Class. Quantum Gravity; 2018; 35, 095016. [DOI: https://dx.doi.org/10.1088/1361-6382/aab793]
40. Vajente, G.; Huang, Y.; Isi, M.; Driggers, J.C.; Kissel, J.S.; Szczepańczyk, M.; Vitale, S. Machine-learning nonstationary noise out of gravitational-wave detectors. Phys. Rev.; 2020; 101, 042003. [DOI: https://dx.doi.org/10.1103/PhysRevD.101.042003]
41. Rezaei, S.; Chegeni, A.; Javadpour, A.; VafaeiSadr, A.; Cao, L.; Röttgering, H.; Staring, M. Bridging gaps with computer vision: AI in (bio)medical imaging and astronomy. Astron. Comput.; 2025; 51, 100921. [DOI: https://dx.doi.org/10.1016/j.ascom.2024.100921]
42. Baron, D. Machine learning in astronomy: A practical overview. arXiv; 2019; arXiv: 1904.07248
43. Ntampaka, M.; Avestruz, C.; Boada, S.; Caldeira, J.; Cisewski-Kehe, J.; Di Stefano, R.; Dvorkin, C.; Evrard, A.E.; Farahi, A.; Finkbeiner, D. et al. The role of machine learning in the next decade of cosmology. arXiv; 2019; arXiv: 1902.10159
44. Croom, S.M.; Lawrence, J.S.; Bland-Hawthorn, J.; Bryant, J.J.; Fogarty, L.; Richards, S.; Goodwin, M.; Farrell, T.; Miziarski, S.; Heald, R. et al. The Sydney-AAO multi-object integral field spectrograph. Mon. Not. R. Astron. Soc.; 2012; 421, pp. 872-893. [DOI: https://dx.doi.org/10.1111/j.1365-2966.2011.20365.x]
45. Bryant, J.J.; Owers, M.S.; Robotham, A.; Croom, S.M.; Driver, S.P.; Drinkwater, M.J.; Lorente, N.P.; Cortese, L.; Scott, N.; Colless, M. et al. The SAMI Galaxy Survey: Instrument specification and target selection. Mon. Not. R. Astron. Soc.; 2015; 447, pp. 2857-2879. [DOI: https://dx.doi.org/10.1093/mnras/stu2635]
46. Sharp, R.; Saunders, W.; Smith, G.; Churilov, V.; Correll, D.; Dawson, J.; Farrel, T.; Frost, G.; Haynes, R.; Heald, R. et al. Performance of AAOmega: The AAT multi-purpose fiber-fed spectrograph. Proceedings of the Ground-Based and Airborne Instrumentation for Astronomy; Orlando, FL, USA, 24–31 May 2006; Volume 6269, pp. 152-164.
47. Sharp, R.; Allen, J.T.; Fogarty, L.M.; Croom, S.M.; Cortese, L.; Green, A.W.; Nielsen, J.; Richards, S.; Scott, N.; Taylor, E.N. et al. The SAMI Galaxy Survey: Cubism and covariance, putting round pegs into square holes. Mon. Not. R. Astron. Soc.; 2015; 446, pp. 1551-1566. [DOI: https://dx.doi.org/10.1093/mnras/stu2055]
48. Bland-Hawthorn, J.; Bryant, J.; Robertson, G.; Gillingham, P.; O’Byrne, J.; Cecil, G.; Haynes, R.; Croom, S.; Ellis, S.; Maack, M. et al. Hexabundles: Imaging fiber arrays for low-light astronomical applications. Opt. Express; 2011; 19, pp. 2649-2661. [DOI: https://dx.doi.org/10.1364/OE.19.002649]
49. Allen, J.T.; Croom, S.M.; Konstantopoulos, I.S.; Bryant, J.J.; Sharp, R.; Cecil, G.; Fogarty, L.M.; Foster, C.; Green, A.W.; Ho, I.T. et al. The SAMI Galaxy Survey: Early data release. Mon. Not. R. Astron. Soc.; 2015; 446, pp. 1567-1583. [DOI: https://dx.doi.org/10.1093/mnras/stu2057]
50. Green, A.W.; Croom, S.M.; Scott, N.; Cortese, L.; Medling, A.M.; D’eugenio, F.; Bryant, J.J.; Bland-Hawthorn, J.; Allen, J.T.; Sharp, R. et al. The SAMI Galaxy Survey: Data Release One with emission-line physics value-added products. Mon. Not. R. Astron. Soc.; 2018; 475, pp. 716-734. [DOI: https://dx.doi.org/10.1093/mnras/stx3135]
51. Scott, N.; van de Sande, J.; Croom, S.M.; Groves, B.; Owers, M.S.; Poetrodjojo, H.; D’Eugenio, F.; Medling, A.M.; Barat, D.; Barone, T.M. et al. The SAMI Galaxy Survey: Data Release Two with absorption-line physics value-added products. Mon. Not. R. Astron. Soc.; 2018; 481, pp. 2299-2319. [DOI: https://dx.doi.org/10.1093/mnras/sty2355]
52. Van De Sande, J.; Bland-Hawthorn, J.; Fogarty, L.M.; Cortese, L.; d’Eugenio, F.; Croom, S.M.; Scott, N.; Allen, J.T.; Brough, S.; Bryant, J.J. et al. The SAMI Galaxy Survey: Revisiting galaxy classification through high-order stellar kinematics. Astrophys. J.; 2017; 835, 104. [DOI: https://dx.doi.org/10.3847/1538-4357/835/1/104]
53. Emsellem, E.; Cappellari, M.; Krajnović, D.; Alatalo, K.; Blitz, L.; Bois, M.; Bournaud, F.; Bureau, M.; Davies, R.L.; Davis, T.A. et al. The ATLAS3D project—III. A census of the stellar angular momentum within the effective radius of early-type galaxies: Unveiling the distribution of fast and slow rotators. Mon. Not. R. Astron. Soc.; 2011; 414, pp. 888-912. [DOI: https://dx.doi.org/10.1111/j.1365-2966.2011.18496.x]
54. Zhao, Z.Q.; Zheng, P.; Xu, S.t.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst.; 2019; 30, pp. 3212-3232. [DOI: https://dx.doi.org/10.1109/TNNLS.2018.2876865]
55. Wersing, H.; Körner, E. Learning optimized features for hierarchical models of invariant object recognition. Neural Comput.; 2003; 15, pp. 1559-1588. [DOI: https://dx.doi.org/10.1162/089976603321891800]
56. O’Malley, T.; Bursztein, E.; Long, J.; Chollet, F.; Jin, H.; Invernizzi, L.; Jin, H. KerasTuner. 2019; Available online: https://github.com/keras-team/keras-tuner (accessed on 29 January 2025).
57. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv; 2014; arXiv: 1412.6980
58. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. Proceedings of the International Conference on Machine Learning; Honolulu, HI, USA, 23–29 July 2023; pp. 23803-23828.
59. Liang, J. Confusion matrix: Machine learning. Pogil Act. Clgh.; 2022; 3, 4.Available online: https://pac.pogil.org/index.php/pac/article/view/304 (accessed on 12 December 2022).
60. Hanley, J.A. Receiver operating characteristic (ROC) methodology: The state of the art. Crit. Rev. Diagn. Imaging; 1989; 29, pp. 307-335.
61. Harborne, K.E.; Power, C.; Robotham, A.S.G.; Cortese, L.; Taranu, D.S. A numerical twist on the spin parameter, λR. Mon. Not. R. Astron. Soc.; 2019; 483, pp. 249-262. [DOI: https://dx.doi.org/10.1093/mnras/sty3120]
62. Moran, P.A. The interpretation of statistical maps. J. R. Stat. Soc. Ser. (Methodol.); 1948; 10, pp. 243-251. [DOI: https://dx.doi.org/10.1111/j.2517-6161.1948.tb00012.x]
63. Chen, Y. New approaches for calculating Moran’s index of spatial autocorrelation. PLoS ONE; 2013; 8, e68336. [DOI: https://dx.doi.org/10.1371/journal.pone.0068336]
64. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic attribution for deep networks. Proceedings of the International Conference on Machine Learning, PMLR; Sydney, Australia, 6–11 August 2017; pp. 3319-3328.
65. Wilde, J.; Serjeant, S.; Bromley, J.M.; Dickinson, H.; Koopmans, L.V.; Metcalf, R.B. Detecting gravitational lenses using machine learning: Exploring interpretability and sensitivity to rare lensing configurations. Mon. Not. R. Astron. Soc.; 2022; 512, pp. 3464-3479. [DOI: https://dx.doi.org/10.1093/mnras/stac562]
66. Matilla, J.M.Z.; Sharma, M.; Hsu, D.; Haiman, Z. Interpreting deep learning models for weak lensing. Phys. Rev.; 2020; 102, 123506. [DOI: https://dx.doi.org/10.1103/PhysRevD.102.123506]
67. Cordonnier, J.B.; Loukas, A.; Jaggi, M. On the relationship between self-attention and convolutional layers. arXiv; 2019; arXiv: 1911.03584
68. Breiman, L. Random forests. Mach. Learn.; 2001; 45, pp. 5-32. [DOI: https://dx.doi.org/10.1023/A:1010933404324]
69. Suthaharan, S.; Suthaharan, S. Support vector machine. Machine Learning Models and Algorithms for Big Data Classification: Thinking with Examples for Effective Learning; Springer: Boston, MA, USA, 2016; pp. 207-235.
70. Webb, G.I.; Keogh, E.; Miikkulainen, R. Naïve Bayes. Encycl. Mach. Learn.; 2010; 15, pp. 713-714.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Distinguishing galaxies as either fast or slow rotators plays a vital role in understanding the processes behind galaxy formation and evolution. Standard techniques, which are based on the
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details



1 Dipartimento di Fisica e Astronomia “G. Galilei”, Università di Padova, Via Marzolo 8, 35131 Padova, Italy;
2 Leiden Institution of Advance Computer Science, P.O. Box 9513, 2300 RA Leiden, The Netherlands;
3 ILEWG LUNEX-EuroMoonMars Earth-Space, ESTEC European Space Agency, Keplerlaan 1, 2201 AZ Noordwijk, The Netherlands;
4 ILEWG LUNEX-EuroMoonMars Earth-Space, ESTEC European Space Agency, Keplerlaan 1, 2201 AZ Noordwijk, The Netherlands;
5 Leiden Institution of Advance Computer Science, P.O. Box 9513, 2300 RA Leiden, The Netherlands;