Page 711 - AI for Good Innovate for Impact
P. 711
AI for Good Innovate for Impact
[8] Huang, Y. (2025). SpecCLIP: Combining and translating spectroscopic measurements
for stars. In Proceedings of IAU Symposium 397: UniversAI: Exploring the Universe with
Artificial Intelligence .
[9] Tao, Y., & Zuo, X. (2025). FALCO: a Foundation model of Astronomical Light Curves for
time dOmain astronomy. In Proceedings of IAU Symposium 397: UniversAI: Exploring cities 4.8: Smart home/
the Universe with Artificial Intelligence.
[10] Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell,
[]
A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. Learning Transferable Visual Models
From Natural Language Supervision. Proceedings of the 38th International Conference
on Machine Learning (ICML), 2021.
[11] Zhang, C., Shen, Y., Zhou, Y., & Tang, J. AstroCLIP: Contrastive Language-Image
Pretraining for Astronomy. arXiv preprint, arXiv:2311.15756, 2023.
[12] An Yang, Baosong Yang, Beichen Zhang,et al., Qwen2.5 Technical Report. arXiv preprint,
arXiv:2412.15115 2024.
[13] Weber, M., Fu, D., Anthony, Q., Oren, Y., Adams, S., Alexandrov, A., ... & Zhang, C.
(2024). RedPajama: an open dataset for training large language models arXiv preprint,
arXiv:2411.12372.
[14] Soldaini, L., Kinney, R., Bhagia, A., Schwenk, D., Atkinson, D., Authur, R., ... & Hajishirzi,
H. (2024). Dolma: Data and tools for generating and inspecting OLMo pre-training data,
arXiv preprint, arXiv:2402.00159.
[15] https:// huggingface .co/ datasets/ anon8231489123/ ShareGPT _Vicuna _unfiltered .
[16] https:// huggingface .co/ datasets/ meta -math/ MetaMathQA .
[17] A 72-billion-parameter instruction-tuned model optimized for instruction following and
advanced reasoning from Qwen, https:// huggingface .co/ Qwen/ Qwen2 .5 -72B -Instruct
[18] OpenAI's flagship multimodal model with advanced reasoning and broad domain
coverage, https:// platform .openai .com/ docs/ models/ gpt -4o
[19] DeepSeek's R1 is a large language model with strong reasoning skills, https:// huggingface
.co/ deepseek -ai/ DeepSeek -R1
[20] Tijmen de Haan, Yuan-Sen Ting, Tirthankar Ghosal,et al., AstroMLab 4: Benchmark-
Topping Performance in Astronomy Q&A with a 70B-Parameter Domain-Specialized
Reasoning Model, arXiv preprint, arXiv:2505.17592 2025
[21] AstroMLab's latest Llama-based Model for astronomy domain, https:// huggingface .co/
AstroMLab/ AstroSage -70B
[22] A graduate-level question answering dataset covering a wide range of scientific and
general knowledge domains, https:// huggingface .co/ datasets/ Idavidrein/ gpqa
[23] Wu, J., Liu, P., Deng, J., et al. InternVL-Chat: Vision-Language Foundation Model for
Science with Scientific Multimodal Instruction Tuning. arXiv preprint arXiv:2312.01908,
2023.
675

