Skip to main content

Knowledge Distillation

Training a smaller 'student' model to mimic a larger 'teacher' model, creating more efficient models with similar performance.

Related Terms

Explore More Terms

Browse Full Glossary