For a fistful of dollars: fine-tune LLaMA 2 7B with QLoRA

Julien Simon
Oct 10, 2023

Fine-tuning large language models doesn’t have to be complicated and expensive.

In this tutorial, I provide a step-by-step demonstration of the fine-tuning process for a LLaMA 2 7-billion parameter model. Thanks to LoRA, 4-bit quantization and a modest AWS GPU instance (g5.xlarge), total cost is just a fistful of dollars 🤠 🤠 🤠

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

Julien Simon
Julien Simon

No responses yet