Papers

Vision Transformers for Cosmological Fields Application to Weak Lensing Mass Maps

J Kakadia, S Agrawal, K Zhong, B Jain - arXiv preprint arXiv:2512.07125, 2025 - arxiv.org
Astrophysics paper astro-ph.CO Suggest

… In Section 2, we introduce simulation-based inference, how vision models can be used to constrain cosmology from the field level, the simulations we use, and how we …

Link to paper

BibTeX

@article{2512.07125v1,
Author = {Jash Kakadia and Shubh Agrawal and Kunhao Zhong and Bhuvnesh Jain},
Title = {Vision Transformers for Cosmological Fields: Application to Weak Lensing Mass Maps},
Eprint = {2512.07125v1},
ArchivePrefix = {arXiv},
PrimaryClass = {astro-ph.CO},
Abstract = {Weak gravitational lensing is a powerful probe of the universe's growth history. While traditional two-point statistics capture only the Gaussian features of the convergence field, deep learning methods such as convolutional neural networks (CNNs) have shown promise in extracting non-Gaussian information from small-scale, nonlinear structures. In this work, we evaluate the effectiveness of attention-based architectures, including variants of vision transformers (ViTs) and shifted window (Swin) transformers, in constraining the cosmological parameters $Ω_m$ and $S_8$ from weak lensing mass maps. Using a simulation-based inference (SBI) framework, we compare transformer-based methods to CNNs. We also examine performance scaling with the number of available $N$-body simulations, highlighting the importance of pre-training for transformer architectures. We find that the Swin transformer performs significantly better than vanilla ViTs, especially with limited training data. Despite their higher representational capacity, the Figure of Merit for cosmology achieved by transformers is comparable to that of CNNs under realistic noise conditions.},
Year = {2025},
Month = {Dec},
Url = {http://arxiv.org/abs/2512.07125v1},
File = {2512.07125v1.pdf}
}

Share