Google Colab Notebook to Step by Step finetune BLOOMZ-3B #LLM using #LoRA
Model: I fine tuned BLOOMZ-3B, a large language model trained by the BigScience team. The full paper is available here!
Dataset: I used lamini_docs question answering dataset from Hugging Face.
Training Infra: I have used Google Colab T4 runtime with 16GB RAM, so failed to do Full Fine Tuning.
Failed to perform Full Fine Tuning in the Infra: To load 3B parameter model in BFLOAT16 requires 6GB memory for inference & 120GB memory for training.
LoRA: Fine Tuned using LoRA and achieve good results in 3 epochs, as the universal model already has a good language understanding.