Deep Dive: Model Distillation with DistillKit

Julien Simon
Jan 3, 2025

In this deep dive video, we zoom in on model distillation, an advanced technique to build high-performance small language models at a reasonable cost.

We first explain what a model distillation is. Then, we introduce two popular techniques for distillation, logits distillation, and hidden states distillation. We study in detail how they work, and how they’re implemented in the Arcee DistillKit open-source library. Finally, we look at two Arcee models built with distillation, Arcee SuperNova 70B and Arcee SuperNova Medius 14B.

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

Julien Simon
Julien Simon

No responses yet

What are your thoughts?