PAPERCLIP: Associating Astronomical Observations and Natural Language with Multi-Modal Models (2403.08851v1)
Abstract: We present PAPERCLIP (Proposal Abstracts Provide an Effective Representation for Contrastive Language-Image Pre-training), a method which associates astronomical observations imaged by telescopes with natural language using a neural network model. The model is fine-tuned from a pre-trained Contrastive Language-Image Pre-training (CLIP) model using successful observing proposal abstracts and corresponding downstream observations, with the abstracts optionally summarized via guided generation using LLMs. Using observations from the Hubble Space Telescope (HST) as an example, we show that the fine-tuned model embodies a meaningful joint representation between observations and natural language through tests targeting image retrieval (i.e., finding the most relevant observations using natural language queries) and description retrieval (i.e., querying for astrophysical object classes and use cases most relevant to a given observation). Our study demonstrates the potential for using generalist foundation models rather than task-specific models for interacting with astronomical data by leveraging text as an interface.
- Data compression and inference in cosmology with self-supervised machine learning. Monthly Notices of the Royal Astronomical Society, 527(3):7459–7481, 2024.
- The DeepMind JAX Ecosystem, 2020. URL http://github.com/deepmind.
- A foundation model for atomistic materials chemistry. arXiv preprint arXiv:2401.00096, 2023.
- Lukas Biewald. Experiment tracking with weights and biases, 2020. URL https://www.wandb.com/. Software available from wandb.com.
- OmniJet-α𝛼\alphaitalic_α: The first cross-task foundation model for particle physics. 3 2024.
- On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
- A new task: Deriving semantic class targets for the physical sciences. arXiv preprint arXiv:2210.14760, 2022.
- Radio galaxy zoo emu: towards a semantic radio galaxy morphology taxonomy. Monthly Notices of the Royal Astronomical Society, 522(2):2584–2600, 2023.
- JAX: composable transformations of Python+NumPy programs, 2018. URL http://github.com/google/jax.
- Geoclip: Clip-inspired alignment between locations and images for effective worldwide geo-localization. arXiv preprint arXiv:2309.16020, 2023.
- Llm.int8(): 8-bit matrix multiplication for transformers at scale. arXiv preprint arXiv:2208.07339, 2022.
- An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- astroquery: An Astronomical Web-querying Package in Python. Astrophysical Journal, 157:98, March 2019. doi: 10.3847/1538-3881/aafc33.
- Array programming with NumPy. Nature, 585(7825):357–362, September 2020. doi: 10.1038/s41586-020-2649-2. URL https://doi.org/10.1038/s41586-020-2649-2.
- Morpheus: A deep learning framework for the pixel-level analysis of astronomical image data. The Astrophysical Journal Supplement Series, 248(1):20, 2020.
- Estimating galactic distances from images using self-supervised representation learning. arXiv preprint arXiv:2101.04293, 2021a.
- Self-supervised representation learning for astronomical images. The Astrophysical Journal Letters, 911(2):L33, 2021b.
- Flax: A neural network library and ecosystem for JAX, 2023. URL http://github.com/google/flax.
- Masked particle modeling on sets: Towards self-supervised high energy physics foundation models. arXiv preprint arXiv:2401.13537, 2024.
- The dawes review 10: The impact of deep learning for the analysis of galaxy surveys. arXiv preprint arXiv:2210.01813, 2022.
- A brief review of contrastive learning applied to astrophysics. RAS Techniques and Instruments, 2(1):441–452, 2023.
- J. D. Hunter. Matplotlib: A 2d graphics environment. Computing in Science & Engineering, 9(3):90–95, 2007. doi: 10.1109/MCSE.2007.55.
- Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
- Adam: A method for stochastic optimization. In Yoshua Bengio and Yann LeCun (eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. URL http://arxiv.org/abs/1412.6980.
- Jupyter notebooks – a publishing format for reproducible computational workflows. In F. Loizides and B. Schmidt (eds.), Positioning and Power in Academic Publishing: Players, Agents and Agendas, pp. 87 – 90. IOS Press, 2016.
- Astroclip: Cross-modal pre-training for astronomical foundation models. arXiv preprint arXiv:2310.03024, 2023.
- The phangs-hst survey: Physics at high angular resolution in nearby galaxies with the hubble space telescope. The Astrophysical Journal Supplement Series, 258(1):10, 2022.
- A text-guided protein design framework. arXiv preprint arXiv:2302.04611, 2023.
- Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. URL https://openreview.net/forum?id=Bkg6RiCqY7.
- Multiple physics pretraining for physical surrogate models. arXiv preprint arXiv:2310.02994, 2023.
- Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
- Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Cloome: contrastive learning unlocks bioimaging databases for queries with chemical structures. bioRxiv, 2023. doi: 10.1101/2022.11.17.516915. URL https://www.biorxiv.org/content/early/2023/06/01/2022.11.17.516915.
- The cosmic evolution survey (cosmos): overview. The Astrophysical Journal Supplement Series, 172(1):1, 2007.
- Radio galaxy zoo: towards building the first multipurpose foundation model for radio astronomy with self-supervised learning. RAS Techniques and Instruments, 3(1):19–32, 2024.
- Learning useful representations for radio astronomy" in the wild" with contrastive learning. arXiv preprint arXiv:2207.08666, 2022.
- Self-supervised similarity search for large scientific datasets. arXiv preprint arXiv:2110.13151, 2021.
- Mining for strong gravitational lenses with self-supervised learning. The Astrophysical Journal, 932(2):107, 2022.
- Towards foundation models for scientific machine learning: Characterizing scaling and transfer behavior. arXiv preprint arXiv:2306.00258, 2023.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Finetuning foundation models for joint analysis optimization. arXiv preprint arXiv:2401.13536, 2024.
- SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272, 2020. doi: 10.1038/s41592-019-0686-2.
- Rare galaxy classes identified in foundation model representations. arXiv preprint arXiv:2312.02910, 2023.
- Deep transfer learning for star cluster classification: I. application to the phangs–hst survey. Monthly Notices of the Royal Astronomical Society, 493(3):3178–3193, 2020.
- Efficient guided generation for llms. arXiv preprint arXiv:2307.09702, 2023.
- Huggingface’s transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771, 2019.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.