TY - JOUR
T1 - Transfer learning with spinally shared layers
AU - Kabir, H. M.Dipu
AU - Mondal, Subrota Kumar
AU - Alam, Syed Bahauddin
AU - Acharya, U. Rajendra
N1 - Publisher Copyright:
© 2024 The Author(s)
PY - 2024/9
Y1 - 2024/9
N2 - Transfer-learned models have achieved promising performance in numerous fields. However, high-performing transfer-learned models contain a large number of parameters. In this paper, we propose a transfer learning approach with parameter reduction and potential high performance. Although the high performance depends on the nature of the dataset, we ensure the parameter reduction. In the proposed SpinalNet shared parameters, all intermediate-split-incoming parameters except the first-intermediate-split contain a shared value. Therefore, the SpinalNet shared parameters network contains three parameter groups: (1) first input-split to intermediate-split parameters, (2) shared intermediate-split-incoming parameters, and (3) intermediate-split-to-output-split parameters. The total number of parameters becomes lower than the SpinalNet and traditional fully connected layers due to parameter sharing. Besides the overall accuracy, this paper compares the precision, recall, and F1-score of each class as performance criteria. As a result, both parameter reduction and potential performance improvement become possible for the ResNet-type models, VGG-type traditional models, and Vision Transformers. We applied the proposed model to MNIST, STL-10, and COVID-19 datasets to validate our claims. We also provided a posterior plot of the sample from different models for medical practitioners to understand the uncertainty. Example model training scripts of the proposed model are also shared to GitHub.
AB - Transfer-learned models have achieved promising performance in numerous fields. However, high-performing transfer-learned models contain a large number of parameters. In this paper, we propose a transfer learning approach with parameter reduction and potential high performance. Although the high performance depends on the nature of the dataset, we ensure the parameter reduction. In the proposed SpinalNet shared parameters, all intermediate-split-incoming parameters except the first-intermediate-split contain a shared value. Therefore, the SpinalNet shared parameters network contains three parameter groups: (1) first input-split to intermediate-split parameters, (2) shared intermediate-split-incoming parameters, and (3) intermediate-split-to-output-split parameters. The total number of parameters becomes lower than the SpinalNet and traditional fully connected layers due to parameter sharing. Besides the overall accuracy, this paper compares the precision, recall, and F1-score of each class as performance criteria. As a result, both parameter reduction and potential performance improvement become possible for the ResNet-type models, VGG-type traditional models, and Vision Transformers. We applied the proposed model to MNIST, STL-10, and COVID-19 datasets to validate our claims. We also provided a posterior plot of the sample from different models for medical practitioners to understand the uncertainty. Example model training scripts of the proposed model are also shared to GitHub.
KW - COVID
KW - ResNet
KW - SpinalNet
KW - Transformer
KW - Uncertainty
KW - VGG
UR - http://www.scopus.com/inward/record.url?scp=85197385655&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85197385655&partnerID=8YFLogxK
U2 - 10.1016/j.asoc.2024.111908
DO - 10.1016/j.asoc.2024.111908
M3 - Article
AN - SCOPUS:85197385655
SN - 1568-4946
VL - 163
JO - Applied Soft Computing
JF - Applied Soft Computing
M1 - 111908
ER -