What is: Teacher-Tutor-Student Knowledge Distillation?
Source | Parser-Free Virtual Try-on via Distilling Appearance Flows |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Teacher-Tutor-Student Knowledge Distillation is a method for image virtual try-on models. It treats fake images produced by the parser-based method as "tutor knowledge", where the artifacts can be corrected by real "teacher knowledge", which is extracted from the real person images in a self-supervised way. Other than using real images as supervisions, knowledge distillation is formulated in the try-on problem as distilling the appearance flows between the person image and the garment image, enabling the finding of dense correspondences between them to produce high-quality results.