Astroconformer: The prospects of analysing stellar light curves with transformer-based deep learning models

被引:2
|
作者
Pan, Jia-Shu [1 ,2 ]
Ting, Yuan-Sen [1 ,3 ,4 ]
Yu, Jie [3 ,5 ,6 ]
机构
[1] Australian Natl Univ, Res Sch Astron & Astrophys, Cotter Rd, Weston, ACT 2611, Australia
[2] Nanjing Univ, Sch Astron & Space Sci, Nanjing 210093, Peoples R China
[3] Australian Natl Univ, Sch Comp, Acton, ACT 2601, Australia
[4] Ohio State Univ, Dept Astron, Columbus, OH 43210 USA
[5] Max Planck Inst Solar Syst Res, Justus von Liebig Weg 3, D-37077 Gottingen, Germany
[6] Heidelberg Inst Theoret Studies HITS gGmbH, Schloss Wolfsbrunnenweg 35, D-69118 Heidelberg, Germany
基金
澳大利亚研究理事会;
关键词
asteroseismology; methods: data analysis; MAGNETIC-FIELDS; ASTEROSEISMOLOGY; STARS; CORES;
D O I
10.1093/mnras/stae068
中图分类号
P1 [天文学];
学科分类号
0704 ;
摘要
Stellar light curves contain valuable information about oscillations and granulation, offering insights into stars' internal structures and evolutionary states. Traditional asteroseismic techniques, primarily focused on power spectral analysis, often overlook the crucial phase information in these light curves. Addressing this gap, recent machine learning applications, particularly those using Convolutional Neural Networks (CNNs), have made strides in inferring stellar properties from light curves. However, CNNs are limited by their localized feature extraction capabilities. In response, we introduce Astroconformer, a Transformer-based deep learning framework, specifically designed to capture long-range dependencies in stellar light curves. Our empirical analysis centres on estimating surface gravity (log g), using a data set derived from single-quarter Kepler light curves with log g values ranging from 0.2 to 4.4. Astroconformer demonstrates superior performance, achieving a root-mean-square-error (RMSE) of 0.017 dex at log g approximate to 3 in data-rich regimes and up to 0.1 dex in sparser areas. This performance surpasses both K-nearest neighbour models and advanced CNNs. Ablation studies highlight the influence of receptive field size on model effectiveness, with larger fields correlating to improved results. Astroconformer also excels in extracting nu(max) with high precision. It achieves less than 2 per cent relative median absolute error for 90-d red giant light curves. Notably, the error remains under 3 per cent for 30-d light curves, whose oscillations are undetectable by a conventional pipeline in 30 per cent cases. Furthermore, the attention mechanisms in Astroconformer align closely with the characteristics of stellar oscillations and granulation observed in light curves.
引用
收藏
页码:5890 / 5903
页数:14
相关论文
共 50 条
  • [31] Multidomain transformer-based deep learning for early detection of network intrusion
    Liu, Jinxin
    Simsek, Murat
    Nogueira, Michele
    Kantarci, Burak
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 6056 - 6061
  • [32] TemproNet: A transformer-based deep learning model for seawater temperature prediction
    Chen, Qiaochuan
    Cai, Candong
    Chen, Yaoran
    Zhou, Xi
    Zhang, Dan
    Peng, Yan
    OCEAN ENGINEERING, 2024, 293
  • [33] Generalizability of transformer-based deep learning for multidimensional turbulent flow data
    Drikakis, Dimitris
    Kokkinakis, Ioannis William
    Fung, Daryl
    Spottswood, S. Michael
    PHYSICS OF FLUIDS, 2024, 36 (02)
  • [34] Deep-ProBind: binding protein prediction with transformer-based deep learning model
    Khan, Salman
    Noor, Sumaiya
    Awan, Hamid Hussain
    Iqbal, Shehryar
    AlQahtani, Salman A.
    Dilshad, Naqqash
    Ahmad, Nijad
    BMC Bioinformatics, 2025, 26 (01)
  • [35] EEG Classification with Transformer-Based Models
    Sun, Jiayao
    Xie, Jin
    Zhou, Huihui
    2021 IEEE 3RD GLOBAL CONFERENCE ON LIFE SCIENCES AND TECHNOLOGIES (IEEE LIFETECH 2021), 2021, : 92 - 93
  • [36] Transformer-based ensemble deep learning model for EEG-based emotion recognition
    Xiaopeng Si
    Dong Huang
    Yulin Sun
    Shudi Huang
    He Huang
    Dong Ming
    Brain Science Advances, 2023, 9 (03) : 210 - 223
  • [37] Transformer-based deep learning method for optimizing ADMET properties of lead compounds
    Yang, Lijuan
    Jin, Chao
    Yang, Guanghui
    Bing, Zhitong
    Huang, Liang
    Niu, Yuzhen
    Yang, Lei
    PHYSICAL CHEMISTRY CHEMICAL PHYSICS, 2023, 25 (03) : 2377 - 2385
  • [38] Transformer-based deep imitation learning for dual-arm robot manipulation
    Kim, Heecheol
    Ohmura, Yoshiyuki
    Kuniyoshi, Yasuo
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 8965 - 8972
  • [39] Identifying suicidal emotions on social media through transformer-based deep learning
    Dheeraj Kodati
    Ramakrishnudu Tene
    Applied Intelligence, 2023, 53 : 11885 - 11917
  • [40] Contextualized medication information extraction using Transformer-based deep learning architectures
    Chen, Aokun
    Yu, Zehao
    Yang, Xi
    Guo, Yi
    Bian, Jiang
    Wu, Yonghui
    JOURNAL OF BIOMEDICAL INFORMATICS, 2023, 142