--- language: - en pipeline_tag: text-generation tags: - llama-3.1 - astronomy - astrophysics - cosmology - arxiv - llama-cpp - gguf-my-repo inference: false base_model: AstroMLab/AstroSage-8B --- # Triangle104/AstroSage-8B-Q6_K-GGUF This model was converted to GGUF format from [`AstroMLab/AstroSage-8B`](https://huggingface.co/AstroMLab/AstroSage-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/AstroMLab/AstroSage-8B) for more details on the model. --- Model details: - https://arxiv.org/abs/2411.09012 AstroSage-Llama-3.1-8B is a domain-specialized natural-language AI assistant tailored for research in astronomy, astrophysics, and cosmology. Trained on the complete collection of astronomy-related arXiv papers from 2007-2024 along with millions of synthetically-generated question-answer pairs and other astronomical literature, AstroSage-Llama-3.1-8B demonstrates excellent proficiency on a wide range of questions. This achievement demonstrates the potential of domain specialization in AI, suggesting that focused training can yield capabilities exceeding those of much larger, general-purpose models. Model Details Base Architecture: Meta-Llama-3.1-8B Base Model: Meta-Llama-3.1-8B Parameters: 8 billion Training Focus: Astronomy, Astrophysics, Cosmology, and Astronomical Instrumentation License: Llama 3.1 Community License Development Process: Continued Pre-training (CPT) on astronomical literature Supervised Fine-tuning (SFT) on QA pairs and instruction sets Model merging with Meta-Llama-3.1-8B-Instruct (75% CPT+SFT / 25% Meta-Instruct) Using the model import torch from transformers import AutoModelForCausalLM, AutoTokenizer # Load the model and tokenizer model = AutoModelForCausalLM.from_pretrained("AstroMLab/AstroSage-8b", device_map="auto") tokenizer = AutoTokenizer.from_pretrained("AstroMLab/AstroSage-8b") # Function to generate a response def generate_response(prompt): inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate( **inputs, max_new_tokens=128, do_sample=True, pad_token_id=tokenizer.eos_token_id, ) response = outputs[0][inputs['input_ids'].shape[-1]:] decoded = tokenizer.decode(response, skip_special_tokens=True) return decoded # Example usage prompt = """ You are an expert in general astrophysics. Your task is to answer the following question: What are the main components of a galaxy? """ response = generate_response(prompt) print(response) Model Improvements and Performance AstroSage-Llama-3.1-8B shows remarkable performance improvements: Model Score (%) AstroSage-Llama-3.1-8B 80.9 GPT-4o 80.4 LLaMA-3.1-8B 73.7 Gemma-2-9B 71.5 Qwen-2.5-7B 70.4 Yi-1.5-9B 68.4 InternLM-2.5-7B 64.5 Mistral-7B-v0.3 63.9 ChatGLM3-6B 50.4 The model demonstrates: - Outperformance of all 8B parameter models Comparable performance to GPT-4o (80.4%) ~1000x more cost-effective than proprietary models 7 percentage-point improvement over base Llama-3.1-8b model Training Data - Continued Pre-training: ~250,000 arXiv preprints (2007-2024) from astro-ph and gr-qc Astronomy-related Wikipedia articles Selected astronomy textbooks Total: 3.3 billion tokens, 19.9 GB plaintext Supervised Fine-tuning: 8.8 million curated QA pairs Filtered Infinity-Instruct-7M dataset Paper summaries and metadata Total: 2.0 billion tokens, 9.8 GB plaintext Intended Use - Curiosity-driven question answering Brainstorming new ideas Astronomical research assistance Educational support in astronomy Literature review and summarization Scientific explanation of concepts Limitations - Training data cutoff: January 2024 As with all LLMs, hallucinations are possible Limited by 8B parameter size for complex reasoning Paper metadata not perfectly memorized Performance primarily validated on multiple-choice questions Primarily trained for use in English Technical Specifications - Architecture: Based on Meta-Llama 3.1 Training Infrastructure: ORNL OLCF Frontier Hosting: Hugging Face Hub (AstroMLab/AstroSage-8B) Ethical Considerations - While this model is designed for scientific use: Should not be used as sole source for critical research decisions Output should be verified against primary sources May reflect biases present in astronomical literature Citation and Contact - Corresponding author: Tijmen de Haan (tijmen dot dehaan at gmail dot com) AstroMLab: astromachinelearninglab at gmail dot com Please cite the AstroMLab 3 paper when referencing this model: @preprint{dehaan2024astromlab3, title={AstroMLab 3: Achieving GPT-4o Level Performance in Astronomy with a Specialized 8B-Parameter Large Language Model}, author={Tijmen de Haan and Yuan-Sen Ting and Tirthankar Ghosal and Tuan Dung Nguyen and Alberto Accomazzi and Azton Wells and Nesar Ramachandra and Rui Pan and Zechang Sun}, year={2024}, eprint={2411.09012}, archivePrefix={arXiv}, primaryClass={astro-ph.IM}, url={https://arxiv.org/abs/2411.09012}, } --- ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Triangle104/AstroSage-8B-Q6_K-GGUF --hf-file astrosage-8b-q6_k.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Triangle104/AstroSage-8B-Q6_K-GGUF --hf-file astrosage-8b-q6_k.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Triangle104/AstroSage-8B-Q6_K-GGUF --hf-file astrosage-8b-q6_k.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Triangle104/AstroSage-8B-Q6_K-GGUF --hf-file astrosage-8b-q6_k.gguf -c 2048 ```