Task-Orientated Feature Distillation

Linfeng Zhang Tsinghua University Yukang Shi Xi’an Jiaotong University Zuoqiang Shi Tsinghua University Kaisheng Ma Tsinghua University Chenglong Bao Tsinghua University

TBD mathscidoc:2206.43016

NeurIPS, 2020.7
Feature distillation, a primary method in knowledge distillation, always leads to significant accuracy improvements. Most existing methods distill features in the teacher network through a manually designed transformation. In this paper, we propose a novel distillation method named task-oriented feature distillation (TOFD) where the transformation is convolutional layers that are trained in a data-driven manner by task loss. As a result, the task-oriented information in the features can be captured and distilled to students. Moreover, an orthogonal loss is applied to the feature resizing layer in TOFD to improve the performance of knowledge distillation. Experiments show that TOFD outperforms other distillation methods by a large margin on both image classification and 3D classification tasks. Codes have been released in Github[3].
No keywords uploaded!
[ Download ] [ 2022-06-15 22:20:40 uploaded by Baocl ] [ 409 downloads ] [ 0 comments ]
@inproceedings{linfeng2020task-orientated,
  title={Task-Orientated Feature Distillation},
  author={Linfeng Zhang, Yukang Shi, Zuoqiang Shi, Kaisheng Ma, and Chenglong Bao},
  url={http://archive.ymsc.tsinghua.edu.cn/pacm_paperurl/20220615222041950422372},
  booktitle={NeurIPS},
  year={2020},
}
Linfeng Zhang, Yukang Shi, Zuoqiang Shi, Kaisheng Ma, and Chenglong Bao. Task-Orientated Feature Distillation. 2020. In NeurIPS. http://archive.ymsc.tsinghua.edu.cn/pacm_paperurl/20220615222041950422372.
Please log in for comment!
 
 
Contact us: office-iccm@tsinghua.edu.cn | Copyright Reserved