ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2309.06126
11
20

AstroLLaMA: Towards Specialized Foundation Models in Astronomy

12 September 2023
Tuan Dung Nguyen
Yuan-Sen Ting
I. Ciucă
Charlie OÑeill
Ze-Chang Sun
Maja Jabłońska
Sandor Kruk
Ernest Perkowski
Jack Miller
Jason Li
J. Peek
K. Iyer
Tomasz Ró.zañski
P. Khetarpal
Sharaf Zaman
D. Brodrick
Sergio J. Rodríguez Méndez
Thang Bui
Alyssa Goodman
Alberto Accomazzi
J. P. Naiman
Jesse Cranney
Kevin Schawinski
UniverseTBD
ArXivPDFHTML
Abstract

Large language models excel in many human-language tasks but often falter in highly specialized domains like scholarly astronomy. To bridge this gap, we introduce AstroLLaMA, a 7-billion-parameter model fine-tuned from LLaMA-2 using over 300,000 astronomy abstracts from arXiv. Optimized for traditional causal language modeling, AstroLLaMA achieves a 30% lower perplexity than Llama-2, showing marked domain adaptation. Our model generates more insightful and scientifically relevant text completions and embedding extraction than state-of-the-arts foundation models despite having significantly fewer parameters. AstroLLaMA serves as a robust, domain-specific model with broad fine-tuning potential. Its public release aims to spur astronomy-focused research, including automatic paper summarization and conversational agent development.

View on arXiv
Comments on this paper