Knowledge Distillation
Knowledge distillation is a technique that transfers the knowledge of a large, powerful AI model (the teacher) to a smaller, faster model (the student), enabling efficient deployment without a proportional loss in quality.
What is Knowledge Distillation?
Why Distillation Matters for Business
Related Terms
Explore further
FAQ
Frequently asked questions
Quality loss varies by task and compression ratio. A well-distilled model typically retains 90-98% of the teacher's performance. For specific, well-defined tasks, distilled models can sometimes match the teacher's quality because they learn to specialise.
This depends on the provider's terms of service. Many providers prohibit using their model outputs to train competing models. However, some allow distillation for internal use or offer distillation as a service. Always check the licence terms.
Quantisation reduces precision of an existing model's weights (simpler, smaller gains). Distillation creates an entirely new, smaller model (more complex, larger gains). They can be combined — distil first, then quantise the resulting model for maximum efficiency.
Need help implementing this?
Our team can help you apply these concepts to your business. Book a free strategy call.