Page 711 - AI for Good Innovate for Impact
P. 711

AI for Good Innovate for Impact



               [8]  Huang, Y. (2025). SpecCLIP: Combining and translating spectroscopic measurements
                    for stars. In Proceedings of IAU Symposium 397: UniversAI: Exploring the Universe with
                    Artificial Intelligence .
               [9]  Tao, Y., & Zuo, X. (2025). FALCO: a Foundation model of Astronomical Light Curves for
                    time dOmain astronomy. In Proceedings of IAU Symposium 397: UniversAI: Exploring               cities  4.8: Smart home/
                    the Universe with Artificial Intelligence.
               [10]   Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell,
                    []
                    A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. Learning Transferable Visual Models
                    From Natural Language Supervision. Proceedings of the 38th International Conference
                    on Machine Learning (ICML), 2021.
               [11]  Zhang, C., Shen, Y., Zhou, Y., & Tang, J. AstroCLIP: Contrastive Language-Image
                    Pretraining for Astronomy. arXiv preprint, arXiv:2311.15756, 2023.
               [12]  An Yang, Baosong Yang, Beichen Zhang,et al., Qwen2.5 Technical Report. arXiv preprint,
                    arXiv:2412.15115 2024.
               [13]  Weber, M., Fu, D., Anthony, Q., Oren, Y., Adams, S., Alexandrov, A., ... & Zhang, C.
                    (2024). RedPajama: an open dataset for training large language models arXiv preprint,
                    arXiv:2411.12372.
               [14]  Soldaini, L., Kinney, R., Bhagia, A., Schwenk, D., Atkinson, D., Authur, R., ... & Hajishirzi,
                    H. (2024). Dolma: Data and tools for generating and inspecting OLMo pre-training data,
                    arXiv preprint, arXiv:2402.00159.
               [15]  https:// huggingface .co/ datasets/ anon8231489123/ ShareGPT _Vicuna _unfiltered .
               [16]  https:// huggingface .co/ datasets/ meta -math/ MetaMathQA .
               [17]  A 72-billion-parameter instruction-tuned model optimized for instruction following and
                    advanced reasoning from Qwen, https:// huggingface .co/ Qwen/ Qwen2 .5 -72B -Instruct
               [18]  OpenAI's flagship multimodal model  with advanced reasoning and broad domain
                    coverage, https:// platform .openai .com/ docs/ models/ gpt -4o
               [19]  DeepSeek's R1 is a large language model with strong reasoning skills, https:// huggingface
                    .co/ deepseek -ai/ DeepSeek -R1
               [20]  Tijmen de Haan, Yuan-Sen Ting, Tirthankar Ghosal,et al., AstroMLab 4: Benchmark-
                    Topping Performance in Astronomy Q&A with a 70B-Parameter Domain-Specialized
                    Reasoning Model, arXiv preprint, arXiv:2505.17592 2025
               [21]  AstroMLab's latest Llama-based Model for astronomy domain, https:// huggingface .co/
                    AstroMLab/ AstroSage -70B
               [22]  A graduate-level question answering dataset covering a wide range of scientific and
                    general knowledge domains, https:// huggingface .co/ datasets/ Idavidrein/ gpqa
               [23]  Wu, J., Liu, P., Deng, J., et al. InternVL-Chat: Vision-Language Foundation Model for
                    Science with Scientific Multimodal Instruction Tuning. arXiv preprint arXiv:2312.01908,
                    2023.



























                                                                                                    675
   706   707   708   709   710   711   712   713   714   715   716