LLM Fine-Tuning
- Tech Stack: Python, Transformers, PEFT, BitsAndBytes, HuggingFace, Weights & Biases (wandb)
- Github URL: Project Link
Fine-tuned a 4-bit quantized pre-trained LLM (Llama 3.2) to summarize conversational text efficiently using LoRA (Low-Rank Adaptation) for parameter-efficient training.
Leveraged Weights and Biases (WandB) for real-time monitoring of training metrics and hyperparameter optimization. Evaluated the model’s summarization performance using the ROUGE score.