Knowledge Distillation
Knowledge Distillation is a model compression technique where a smaller model (student) learns to replicate the behavior of a larger, more complex model (teacher). Instead of relying solely on original training data, the student also learns from the teacher’s output...