Rights statement: This is a pre-copy-editing, author-produced PDF of an article accepted for publication in Monthly Notices of the Royal Astronomical Society following peer review. The definitive publisher-authenticated version Nima Sedaghat, Martino Romaniello, Jonathan E Carrick, François-Xavier Pineau, Machines learn to infer stellar parameters just by looking at a large number of spectra, Monthly Notices of the Royal Astronomical Society, Volume 501, Issue 4, March 2021, Pages 6026–6041, https://doi.org/10.1093/mnras/staa3540 is available online at: https://academic.oup.com/mnras/article-abstract/501/4/6026/6121645
Accepted author manuscript, 3.67 MB, PDF document
Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License
Final published version
Research output: Contribution to Journal/Magazine › Journal article › peer-review
Research output: Contribution to Journal/Magazine › Journal article › peer-review
}
TY - JOUR
T1 - Machines Learn to Infer Stellar Parameters Just by Looking at a Large Number of Spectra
AU - Sedaghat, Nima
AU - Romaniello, Martino
AU - Carrick, Jon
AU - Pineau, François-Xavier
N1 - This is a pre-copy-editing, author-produced PDF of an article accepted for publication in Monthly Notices of the Royal Astronomical Society following peer review. The definitive publisher-authenticated version Nima Sedaghat, Martino Romaniello, Jonathan E Carrick, François-Xavier Pineau, Machines learn to infer stellar parameters just by looking at a large number of spectra, Monthly Notices of the Royal Astronomical Society, Volume 501, Issue 4, March 2021, Pages 6026–6041, https://doi.org/10.1093/mnras/staa3540 is available online at: https://academic.oup.com/mnras/article-abstract/501/4/6026/6121645
PY - 2021/3/1
Y1 - 2021/3/1
N2 - Machine learning has been widely applied to clearly defined problems of astronomy and astrophysics. However, deep learning and its conceptual differences to classical machine learning have been largely overlooked in these fields. The broad hypothesis behind our work is that letting the abundant real astrophysical data speak for itself, with minimal supervision and no labels, can reveal interesting patterns that may facilitate discovery of novel physical relationships. Here, as the first step, we seek to interpret the representations a deep convolutional neural network chooses to learn, and find correlations in them with current physical understanding. We train an encoder–decoder architecture on the self-supervised auxiliary task of reconstruction to allow it to learn general representations without bias towards any specific task. By exerting weak disentanglement at the information bottleneck of the network, we implicitly enforce interpretability in the learned features. We develop two independent statistical and information-theoretical methods for finding the number of learned informative features, as well as measuring their true correlation with astrophysical validation labels. As a case study, we apply this method to a data set of ∼270 000 stellar spectra, each of which comprising ∼300 000 dimensions. We find that the network clearly assigns specific nodes to estimate (notions of) parameters such as radial velocity and effective temperature without being asked to do so, all in a completely physics-agnostic process. This supports the first part of our hypothesis. Moreover, we find with high confidence that there are ∼4 more independently informative dimensions that do not show a direct correlation with our validation parameters, presenting potential room for future studies.
AB - Machine learning has been widely applied to clearly defined problems of astronomy and astrophysics. However, deep learning and its conceptual differences to classical machine learning have been largely overlooked in these fields. The broad hypothesis behind our work is that letting the abundant real astrophysical data speak for itself, with minimal supervision and no labels, can reveal interesting patterns that may facilitate discovery of novel physical relationships. Here, as the first step, we seek to interpret the representations a deep convolutional neural network chooses to learn, and find correlations in them with current physical understanding. We train an encoder–decoder architecture on the self-supervised auxiliary task of reconstruction to allow it to learn general representations without bias towards any specific task. By exerting weak disentanglement at the information bottleneck of the network, we implicitly enforce interpretability in the learned features. We develop two independent statistical and information-theoretical methods for finding the number of learned informative features, as well as measuring their true correlation with astrophysical validation labels. As a case study, we apply this method to a data set of ∼270 000 stellar spectra, each of which comprising ∼300 000 dimensions. We find that the network clearly assigns specific nodes to estimate (notions of) parameters such as radial velocity and effective temperature without being asked to do so, all in a completely physics-agnostic process. This supports the first part of our hypothesis. Moreover, we find with high confidence that there are ∼4 more independently informative dimensions that do not show a direct correlation with our validation parameters, presenting potential room for future studies.
KW - methods: data analysis
KW - methods: numerical
KW - techniques: spectroscopic
U2 - 10.1093/mnras/staa3540
DO - 10.1093/mnras/staa3540
M3 - Journal article
VL - 501
SP - 6026
EP - 6041
JO - Monthly Notices of the Royal Astronomical Society
JF - Monthly Notices of the Royal Astronomical Society
SN - 0035-8711
IS - 4
ER -