Knowledge distillation for image signal processing using only the generator portion of a GAN
SCIE
SCOPUS
- Title
- Knowledge distillation for image signal processing using only the generator portion of a GAN
- Authors
- HEO, YOUNG JUN; LEE, SUNG GU
- Date Issued
- 2022-11
- Publisher
- MDPI AG
- Abstract
- Knowledge distillation, in which the parameter values learned in a large teacher network are transferred to a smaller student network, is a popular and effective network compression method. Recently, researchers have proposed methods to improve the performance of a student network by using a Generative Adverserial Network (GAN). However, because a GAN is an architecture that is ideally used to create realistic synthetic images, a pure GAN architecture may not be ideally suited for knowledge distillation. In knowledge distillation for image signal processing, synthetic images do not need to be realistic, but instead should include features that help the training of the student network. In the proposed Generative Image Processing (GIP) method, this is accomplished by using only the generator portion of a GAN and utilizing special techniques to capture the distinguishing feature capability of the teacher network. Experimental results show that the GIP method outperforms knowledge distillation using GANs as well as training using only knowledge distillation.
- URI
- https://oasis.postech.ac.kr/handle/2014.oak/114630
- DOI
- 10.3390/electronics11223815
- ISSN
- 2079-9292
- Article Type
- Article
- Citation
- Electronics (Basel), vol. 11, no. 22, 2022-11
- Files in This Item:
- There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.