Loading...
Loading...
A specialized training pipeline for fine-tuning Google's Gemma 3 (1B) model on astronomy MCQ datasets using Parameter-Efficient Fine-Tuning (PEFT).
This project explores the capacity of Small Language Models (SLMs) to handle domain-specific reasoning. By fine-tuning Gemma 3 using LoRA, the system transforms general-purpose weights into an astronomy expert capable of answering complex multiple-choice questions while maintaining a low computational footprint.