• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Chen, Weijie (Chen, Weijie.) [1] | Xuan, Yunyi (Xuan, Yunyi.) [2] | Yang, Shicai (Yang, Shicai.) [3] | Xie, Di (Xie, Di.) [4] | Lin, Luojun (Lin, Luojun.) [5] (Scholars:林洛君) | Zhuang, Yueting (Zhuang, Yueting.) [6]

Indexed by:

EI Scopus SCIE

Abstract:

Data-Free Knowledge Distillation (DFKD) aims to craft a customized student model from a pre-trained teacher model by synthesizing surrogate training images. However, a seldom-investigated scenario is to distill the knowledge to multiple heterogeneous students simultaneously. In this paper, we aim to study how to improve the performance by coevolving peer students, termed Data-Free Multi-Student Coevolved Distillation (DF-MSCD). Based on previous DFKD methods, we advance DF-MSCD by improving the data quality from the perspective of synthesizing unbiased, informative and diverse surrogate samples: 1) Unbiased. The disconnection of image synthesis among different timestamps during DFKD will lead to an unnoticed class imbalance problem. To tackle this problem, we reform the prior art into an unbiased variant by bridging the label distribution of the synthesized data among different timestamps. 2) Informative. Different from single-student DFKD, we encourage the interactions not only between teacher-student pairs, but also within peer students, driving a more comprehensive knowledge distillation. To this end, we devise a novel Inter-Student Adversarial Learning method to coevolve peer students with mutual benefits. 3) Diverse. To further promote Inter-Student Adversarial Learning, we develop Mixture-of-Generators, in which multiple generators are optimized to synthesize different yet complementary samples by playing min-max games with multiple students. Experiments are conducted to validate the effectiveness and efficiency of the proposed DF-MSCD, surpassing the existing state-of-the-arts on multiple popular benchmarks. To emphasize, our method can obtain heterogeneous students by training once, which is superior to single-student DFKD methods in terms of both training time and testing accuracy.

Keyword:

Adversarial training Knowledge distillation Model inversion Mutual learning Surrogate images

Community:

  • [ 1 ] [Chen, Weijie]Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
  • [ 2 ] [Zhuang, Yueting]Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
  • [ 3 ] [Chen, Weijie]Hikvis Res Inst, Hangzhou, Peoples R China
  • [ 4 ] [Xuan, Yunyi]Hikvis Res Inst, Hangzhou, Peoples R China
  • [ 5 ] [Yang, Shicai]Hikvis Res Inst, Hangzhou, Peoples R China
  • [ 6 ] [Xie, Di]Hikvis Res Inst, Hangzhou, Peoples R China
  • [ 7 ] [Lin, Luojun]Fuzhou Univ, Coll Comp & Data Sci, Fuzhou, Peoples R China

Reprint 's Address:

  • [Zhuang, Yueting]Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China;;[Lin, Luojun]Fuzhou Univ, Coll Comp & Data Sci, Fuzhou, Peoples R China;;

Show more details

Related Keywords:

Source :

KNOWLEDGE-BASED SYSTEMS

ISSN: 0950-7051

Year: 2023

Volume: 283

7 . 2

JCR@2023

7 . 2 0 0

JCR@2023

JCR Journal Grade:1

CAS Journal Grade:2

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 2

Online/Total:57/10367041
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1