Is One Teacher Model Enough to Transfer Knowledge to a Student Model?

Nowadays, the transfer learning technique can be successfully applied in the deep learning field through techniques that fine-tune the CNN’s starting point so it may learn over a huge dataset such as ImageNet and continue to learn on a fixed dataset to achieve better performance. In this paper, we d...

Full description

Saved in:
Bibliographic Details
Main Authors: Nicola Landro, Ignazio Gallo, Riccardo La Grassa
Format: article
Language:EN
Published: MDPI AG 2021
Subjects:
Online Access:https://doaj.org/article/b5508f2d11a04b6db13a9ff530052572
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Nowadays, the transfer learning technique can be successfully applied in the deep learning field through techniques that fine-tune the CNN’s starting point so it may learn over a huge dataset such as ImageNet and continue to learn on a fixed dataset to achieve better performance. In this paper, we designed a transfer learning methodology that combines the learned features of different teachers to a student network in an end-to-end model, improving the performance of the student network in classification tasks over different datasets. In addition to this, we tried to answer the following questions which are in any case directly related to the transfer learning problem addressed here. Is it possible to improve the performance of a small neural network by using the knowledge gained from a more powerful neural network? Can a deep neural network outperform the teacher using transfer learning? Experimental results suggest that neural networks can transfer their learning to student networks using our proposed architecture, designed to bring to light a new interesting approach for transfer learning techniques. Finally, we provide details of the code and the experimental settings.