Transfer learning with spinally shared layers

H. M.Dipu Kabir, Subrota Kumar Mondal, Syed Bahauddin Alam, U. Rajendra Acharya

Research output: Contribution to journalArticlepeer-review

Abstract

Transfer-learned models have achieved promising performance in numerous fields. However, high-performing transfer-learned models contain a large number of parameters. In this paper, we propose a transfer learning approach with parameter reduction and potential high performance. Although the high performance depends on the nature of the dataset, we ensure the parameter reduction. In the proposed SpinalNet shared parameters, all intermediate-split-incoming parameters except the first-intermediate-split contain a shared value. Therefore, the SpinalNet shared parameters network contains three parameter groups: (1) first input-split to intermediate-split parameters, (2) shared intermediate-split-incoming parameters, and (3) intermediate-split-to-output-split parameters. The total number of parameters becomes lower than the SpinalNet and traditional fully connected layers due to parameter sharing. Besides the overall accuracy, this paper compares the precision, recall, and F1-score of each class as performance criteria. As a result, both parameter reduction and potential performance improvement become possible for the ResNet-type models, VGG-type traditional models, and Vision Transformers. We applied the proposed model to MNIST, STL-10, and COVID-19 datasets to validate our claims. We also provided a posterior plot of the sample from different models for medical practitioners to understand the uncertainty. Example model training scripts of the proposed model are also shared to GitHub.

Original languageEnglish (US)
Article number111908
JournalApplied Soft Computing
Volume163
DOIs
StatePublished - Sep 2024

Keywords

  • COVID
  • ResNet
  • SpinalNet
  • Transformer
  • Uncertainty
  • VGG

ASJC Scopus subject areas

  • Software

Fingerprint

Dive into the research topics of 'Transfer learning with spinally shared layers'. Together they form a unique fingerprint.

Cite this