Astroconformer: The prospects of analysing stellar light curves with transformer-based deep learning models

被引:2
|
作者
Pan, Jia-Shu [1 ,2 ]
Ting, Yuan-Sen [1 ,3 ,4 ]
Yu, Jie [3 ,5 ,6 ]
机构
[1] Australian Natl Univ, Res Sch Astron & Astrophys, Cotter Rd, Weston, ACT 2611, Australia
[2] Nanjing Univ, Sch Astron & Space Sci, Nanjing 210093, Peoples R China
[3] Australian Natl Univ, Sch Comp, Acton, ACT 2601, Australia
[4] Ohio State Univ, Dept Astron, Columbus, OH 43210 USA
[5] Max Planck Inst Solar Syst Res, Justus von Liebig Weg 3, D-37077 Gottingen, Germany
[6] Heidelberg Inst Theoret Studies HITS gGmbH, Schloss Wolfsbrunnenweg 35, D-69118 Heidelberg, Germany
基金
澳大利亚研究理事会;
关键词
asteroseismology; methods: data analysis; MAGNETIC-FIELDS; ASTEROSEISMOLOGY; STARS; CORES;
D O I
10.1093/mnras/stae068
中图分类号
P1 [天文学];
学科分类号
0704 ;
摘要
Stellar light curves contain valuable information about oscillations and granulation, offering insights into stars' internal structures and evolutionary states. Traditional asteroseismic techniques, primarily focused on power spectral analysis, often overlook the crucial phase information in these light curves. Addressing this gap, recent machine learning applications, particularly those using Convolutional Neural Networks (CNNs), have made strides in inferring stellar properties from light curves. However, CNNs are limited by their localized feature extraction capabilities. In response, we introduce Astroconformer, a Transformer-based deep learning framework, specifically designed to capture long-range dependencies in stellar light curves. Our empirical analysis centres on estimating surface gravity (log g), using a data set derived from single-quarter Kepler light curves with log g values ranging from 0.2 to 4.4. Astroconformer demonstrates superior performance, achieving a root-mean-square-error (RMSE) of 0.017 dex at log g approximate to 3 in data-rich regimes and up to 0.1 dex in sparser areas. This performance surpasses both K-nearest neighbour models and advanced CNNs. Ablation studies highlight the influence of receptive field size on model effectiveness, with larger fields correlating to improved results. Astroconformer also excels in extracting nu(max) with high precision. It achieves less than 2 per cent relative median absolute error for 90-d red giant light curves. Notably, the error remains under 3 per cent for 30-d light curves, whose oscillations are undetectable by a conventional pipeline in 30 per cent cases. Furthermore, the attention mechanisms in Astroconformer align closely with the characteristics of stellar oscillations and granulation observed in light curves.
引用
收藏
页码:5890 / 5903
页数:14
相关论文
共 50 条
  • [1] Transformer-based deep learning models for predicting permeability of porous media
    Meng, Yinquan
    Jiang, Jianguo
    Wu, Jichun
    Wang, Dong
    ADVANCES IN WATER RESOURCES, 2023, 179
  • [2] ASTROMER A transformer-based embedding for the representation of light curves
    Donoso-Oliva, C.
    Becker, I.
    Protopapas, P.
    Cabrera-Vives, G.
    Vishnu, M.
    Vardhan, H.
    ASTRONOMY & ASTROPHYSICS, 2023, 670
  • [3] Transformer-based Extraction of Deep Image Models
    Battis, Verena
    Penner, Alexander
    2022 IEEE 7TH EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY (EUROS&P 2022), 2022, : 320 - 336
  • [4] A performance analysis of transformer-based deep learning models for Arabic image captioning
    Alsayed, Ashwaq
    Qadah, Thamir M.
    Arif, Muhammad
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (09)
  • [5] Transformer-based deep learning models for the sentiment analysis of social media data
    Kokab, Sayyida Tabinda
    Asghar, Sohail
    Naz, Shehneela
    ARRAY, 2022, 14
  • [6] Comparative Analysis of Traditional Machine Learning and Transformer-based Deep Learning Models for Text Classification
    Aydin, Nazif
    Erdem, Osman Ayhan
    Tekerek, Adem
    JOURNAL OF POLYTECHNIC-POLITEKNIK DERGISI, 2024,
  • [7] Abusive Bangla comments detection on Facebook using transformer-based deep learning models
    Tanjim Taharat Aurpa
    Rifat Sadik
    Md Shoaib Ahmed
    Social Network Analysis and Mining, 2022, 12
  • [8] Dementia Detection using Transformer-Based Deep Learning and Natural Language Processing Models
    Saltz, Ploypaphat
    Lin, Shih Yin
    Cheng, Sunny Chieh
    Si, Dong
    2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 509 - 510
  • [9] Abusive Bangla comments detection on Facebook using transformer-based deep learning models
    Aurpa, Tanjim Taharat
    Sadik, Rifat
    Ahmed, Md Shoaib
    SOCIAL NETWORK ANALYSIS AND MINING, 2022, 12 (01)
  • [10] Transformer-Based Federated Learning Models for Recommendation Systems
    Reddy, M. Sujaykumar
    Karnati, Hemanth
    Sundari, L. Mohana
    IEEE ACCESS, 2024, 12 : 109596 - 109607