For a fistful of dollars: fine-tune LLaMA 2 7B with QLoRA

Julien Simon
Oct 10, 2023

Fine-tuning large language models doesn’t have to be complicated and expensive.

In this tutorial, I provide a step-by-step demonstration of the fine-tuning process for a LLaMA 2 7-billion parameter model. Thanks to LoRA, 4-bit quantization and a modest AWS GPU instance (g5.xlarge), total cost is just a fistful of dollars 🤠 🤠 🤠

--

--