• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Chen, Weijie (Chen, Weijie.) [1] | Xuan, Yunyi (Xuan, Yunyi.) [2] | Yang, Shicai (Yang, Shicai.) [3] | Xie, Di (Xie, Di.) [4] | Lin, Luojun (Lin, Luojun.) [5] | Zhuang, Yueting (Zhuang, Yueting.) [6]

Indexed by:

EI

Abstract:

Data-Free Knowledge Distillation (DFKD) aims to craft a customized student model from a pre-trained teacher model by synthesizing surrogate training images. However, a seldom-investigated scenario is to distill the knowledge to multiple heterogeneous students simultaneously. In this paper, we aim to study how to improve the performance by coevolving peer students, termed Data-Free Multi-Student Coevolved Distillation (DF-MSCD). Based on previous DFKD methods, we advance DF-MSCD by improving the data quality from the perspective of synthesizing unbiased, informative and diverse surrogate samples: 1) Unbiased. The disconnection of image synthesis among different timestamps during DFKD will lead to an unnoticed class imbalance problem. To tackle this problem, we reform the prior art into an unbiased variant by bridging the label distribution of the synthesized data among different timestamps. 2) Informative. Different from single-student DFKD, we encourage the interactions not only between teacher–student pairs, but also within peer students, driving a more comprehensive knowledge distillation. To this end, we devise a novel Inter-Student Adversarial Learning method to coevolve peer students with mutual benefits. 3) Diverse. To further promote Inter-Student Adversarial Learning, we develop Mixture-of-Generators, in which multiple generators are optimized to synthesize different yet complementary samples by playing min–max games with multiple students. Experiments are conducted to validate the effectiveness and efficiency of the proposed DF-MSCD, surpassing the existing state-of-the-arts on multiple popular benchmarks. To emphasize, our method can obtain heterogeneous students by training once, which is superior to single-student DFKD methods in terms of both training time and testing accuracy. © 2023 Elsevier B.V.

Keyword:

Distillation Image processing Learning systems Personnel training Students

Community:

  • [ 1 ] [Chen, Weijie]College of Computer Science and Technology, Zhejiang University, China
  • [ 2 ] [Chen, Weijie]Hikvision Research Institute, China
  • [ 3 ] [Xuan, Yunyi]Hikvision Research Institute, China
  • [ 4 ] [Yang, Shicai]Hikvision Research Institute, China
  • [ 5 ] [Xie, Di]Hikvision Research Institute, China
  • [ 6 ] [Lin, Luojun]College of Computer and Data Science, Fuzhou University, China
  • [ 7 ] [Zhuang, Yueting]College of Computer Science and Technology, Zhejiang University, China

Reprint 's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

Knowledge-Based Systems

ISSN: 0950-7051

Year: 2024

Volume: 283

7 . 2 0 0

JCR@2023

CAS Journal Grade:2

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 2

Affiliated Colleges:

Online/Total:448/10369096
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1