TY - JOUR
T1 - The universe is worth 643 pixels
T2 - convolution neural network and vision transformers for cosmology
AU - Hwang, Se Yeon
AU - Sabiu, Cristiano G.
AU - Park, Inkyu
AU - Hong, Sungwook E.
N1 - Publisher Copyright:
© 2023 IOP Publishing Ltd and Sissa Medialab.
PY - 2023/11/1
Y1 - 2023/11/1
N2 - We present a novel approach for estimating cosmological parameters, Ωm , σ8 , w 0, and one derived parameter, S 8, from 3D lightcone data of dark matter halos in redshift space covering a sky area of 40° × 40° and redshift range of 0.3 < z < 0.8, binned to 643 voxels. Using two deep learning algorithms — Convolutional Neural Network (CNN) and Vision Transformer (ViT) — we compare their performance with the standard two-point correlation (2pcf) function. Our results indicate that CNN yields the best performance, while ViT also demonstrates significant potential in predicting cosmological parameters. By combining the outcomes of Vision Transformer, Convolution Neural Network, and 2pcf, we achieved a substantial reduction in error compared to the 2pcf alone. To better understand the inner workings of the machine learning algorithms, we employed the Grad-CAM method to investigate the sources of essential information in heatmaps of the CNN and ViT. Our findings suggest that the algorithms focus on different parts of the density field and redshift depending on which parameter they are predicting. This proof-of-concept work paves the way for incorporating deep learning methods to estimate cosmological parameters from large-scale structures, potentially leading to tighter constraints and improved understanding of the Universe.
AB - We present a novel approach for estimating cosmological parameters, Ωm , σ8 , w 0, and one derived parameter, S 8, from 3D lightcone data of dark matter halos in redshift space covering a sky area of 40° × 40° and redshift range of 0.3 < z < 0.8, binned to 643 voxels. Using two deep learning algorithms — Convolutional Neural Network (CNN) and Vision Transformer (ViT) — we compare their performance with the standard two-point correlation (2pcf) function. Our results indicate that CNN yields the best performance, while ViT also demonstrates significant potential in predicting cosmological parameters. By combining the outcomes of Vision Transformer, Convolution Neural Network, and 2pcf, we achieved a substantial reduction in error compared to the 2pcf alone. To better understand the inner workings of the machine learning algorithms, we employed the Grad-CAM method to investigate the sources of essential information in heatmaps of the CNN and ViT. Our findings suggest that the algorithms focus on different parts of the density field and redshift depending on which parameter they are predicting. This proof-of-concept work paves the way for incorporating deep learning methods to estimate cosmological parameters from large-scale structures, potentially leading to tighter constraints and improved understanding of the Universe.
KW - Machine learning
KW - cosmological parameters from LSS
KW - dark energy experiments
KW - galaxy clustering
UR - http://www.scopus.com/inward/record.url?scp=85178189522&partnerID=8YFLogxK
U2 - 10.1088/1475-7516/2023/11/075
DO - 10.1088/1475-7516/2023/11/075
M3 - Article
AN - SCOPUS:85178189522
SN - 1475-7516
VL - 11
JO - Journal of Cosmology and Astroparticle Physics
JF - Journal of Cosmology and Astroparticle Physics
IS - 75
M1 - 075
ER -