Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models

Trainable parameters and hyperparameters are critical to the development of a deep learning model. However, the components have typically been studied individually, and most studies have found it difficult to investigate the effects of their combination on model performance. We are interested in exa...

Full description

Bibliographic Details
Published in:2023 4th International Conference on Artificial Intelligence and Data Sciences: Discovering Technological Advancement in Artificial Intelligence and Data Science, AiDAS 2023 - Proceedings
Main Author: Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
Format: Conference paper
Language:English
Published: Institute of Electrical and Electronics Engineers Inc. 2023
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85176568821&doi=10.1109%2fAiDAS60501.2023.10284682&partnerID=40&md5=a781f28bb6214e4666ae3ef9ae8f18fe
id 2-s2.0-85176568821
spelling 2-s2.0-85176568821
Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
2023
2023 4th International Conference on Artificial Intelligence and Data Sciences: Discovering Technological Advancement in Artificial Intelligence and Data Science, AiDAS 2023 - Proceedings


10.1109/AiDAS60501.2023.10284682
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85176568821&doi=10.1109%2fAiDAS60501.2023.10284682&partnerID=40&md5=a781f28bb6214e4666ae3ef9ae8f18fe
Trainable parameters and hyperparameters are critical to the development of a deep learning model. However, the components have typically been studied individually, and most studies have found it difficult to investigate the effects of their combination on model performance. We are interested in examining the correlation between the number of trainable parameters in a deep learning model and its performance metrics under different hyperparameters. Specifically, we want to study the effects of using either the Adam or SGD optimizers at three varying learning rates. We use six pre-trained models whose trainable parameters have been quantitatively defined using two strategies: (1) freezing the convolutional basis with partially trainable weights and (2) training the whole model with most trainable weights to obtain a set of trainable parameters. Our experimental result shows a positive correlation between the trainable parameters and the test accuracy regardless of the level of the learning rate. However, for the generalization of the model, it was not guaranteed that a higher number of trainable parameters would lead to higher accuracy and F1 score. We have shown that the correlation between trainable parameters and model generalization becomes positive by using Adam with the smallest learning rate. © 2023 IEEE.
Institute of Electrical and Electronics Engineers Inc.

English
Conference paper

author Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
spellingShingle Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
author_facet Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
author_sort Ong S.-Q.; Isawasan P.; Nair G.; Salleh K.A.; Yusof U.K.
title Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
title_short Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
title_full Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
title_fullStr Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
title_full_unstemmed Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
title_sort Study on the Correlation of Trainable Parameters and Hyperparameters with the Performance of Deep Learning Models
publishDate 2023
container_title 2023 4th International Conference on Artificial Intelligence and Data Sciences: Discovering Technological Advancement in Artificial Intelligence and Data Science, AiDAS 2023 - Proceedings
container_volume
container_issue
doi_str_mv 10.1109/AiDAS60501.2023.10284682
url https://www.scopus.com/inward/record.uri?eid=2-s2.0-85176568821&doi=10.1109%2fAiDAS60501.2023.10284682&partnerID=40&md5=a781f28bb6214e4666ae3ef9ae8f18fe
description Trainable parameters and hyperparameters are critical to the development of a deep learning model. However, the components have typically been studied individually, and most studies have found it difficult to investigate the effects of their combination on model performance. We are interested in examining the correlation between the number of trainable parameters in a deep learning model and its performance metrics under different hyperparameters. Specifically, we want to study the effects of using either the Adam or SGD optimizers at three varying learning rates. We use six pre-trained models whose trainable parameters have been quantitatively defined using two strategies: (1) freezing the convolutional basis with partially trainable weights and (2) training the whole model with most trainable weights to obtain a set of trainable parameters. Our experimental result shows a positive correlation between the trainable parameters and the test accuracy regardless of the level of the learning rate. However, for the generalization of the model, it was not guaranteed that a higher number of trainable parameters would lead to higher accuracy and F1 score. We have shown that the correlation between trainable parameters and model generalization becomes positive by using Adam with the smallest learning rate. © 2023 IEEE.
publisher Institute of Electrical and Electronics Engineers Inc.
issn
language English
format Conference paper
accesstype
record_format scopus
collection Scopus
_version_ 1809677588410400768