Multi objective hyperparameter tuning via random search on deep learning models
This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and Ale...
Published in: | Telkomnika (Telecommunication Computing Electronics and Control) |
---|---|
Main Author: | |
Format: | Article |
Language: | English |
Published: |
Universitas Ahmad Dahlan
2024
|
Online Access: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde |
id |
2-s2.0-85197106107 |
---|---|
spelling |
2-s2.0-85197106107 Rom A.R.M.; Jamil N.; Ibrahim S. Multi objective hyperparameter tuning via random search on deep learning models 2024 Telkomnika (Telecommunication Computing Electronics and Control) 22 4 10.12928/TELKOMNIKA.v22i4.25847 https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and AlexNet implemented on prominent benchmark datasets of Modified National Institute of Standards and Technology (MNIST) and Canadian Institute for Advanced Research-10 (CIFAR-10). In the context of this study, the evaluation will be adopting a multi-objective framework, navigating the delicate trade-offs between conflicting performance metrics, including accuracy, F1-score, and model parameter size. The primary objective of employing a multi-objective evaluation framework is to enhance the understanding regarding the interactions of these performance metrics interact and influence each other. In real-world scenarios, DL models often need to strike a balance between these conflicting objectives. This research adds to the increasing wealth of knowledge in hyperparameter tuning for DL models and serves as a reference point for practitioners seeking to optimize their DL architectures. The results of our analysis are positioned to provide invaluable insights into the intricate balancing act required during the process of hyperparameter fine-tuning. These insights will contribute to the ongoing advancement of best practices in optimizing DL models and facilitating the ongoing optimization of the DL models. © (2024), (Universitas Ahmad Dahlan). All rights reserved. Universitas Ahmad Dahlan 16936930 English Article |
author |
Rom A.R.M.; Jamil N.; Ibrahim S. |
spellingShingle |
Rom A.R.M.; Jamil N.; Ibrahim S. Multi objective hyperparameter tuning via random search on deep learning models |
author_facet |
Rom A.R.M.; Jamil N.; Ibrahim S. |
author_sort |
Rom A.R.M.; Jamil N.; Ibrahim S. |
title |
Multi objective hyperparameter tuning via random search on deep learning models |
title_short |
Multi objective hyperparameter tuning via random search on deep learning models |
title_full |
Multi objective hyperparameter tuning via random search on deep learning models |
title_fullStr |
Multi objective hyperparameter tuning via random search on deep learning models |
title_full_unstemmed |
Multi objective hyperparameter tuning via random search on deep learning models |
title_sort |
Multi objective hyperparameter tuning via random search on deep learning models |
publishDate |
2024 |
container_title |
Telkomnika (Telecommunication Computing Electronics and Control) |
container_volume |
22 |
container_issue |
4 |
doi_str_mv |
10.12928/TELKOMNIKA.v22i4.25847 |
url |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde |
description |
This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and AlexNet implemented on prominent benchmark datasets of Modified National Institute of Standards and Technology (MNIST) and Canadian Institute for Advanced Research-10 (CIFAR-10). In the context of this study, the evaluation will be adopting a multi-objective framework, navigating the delicate trade-offs between conflicting performance metrics, including accuracy, F1-score, and model parameter size. The primary objective of employing a multi-objective evaluation framework is to enhance the understanding regarding the interactions of these performance metrics interact and influence each other. In real-world scenarios, DL models often need to strike a balance between these conflicting objectives. This research adds to the increasing wealth of knowledge in hyperparameter tuning for DL models and serves as a reference point for practitioners seeking to optimize their DL architectures. The results of our analysis are positioned to provide invaluable insights into the intricate balancing act required during the process of hyperparameter fine-tuning. These insights will contribute to the ongoing advancement of best practices in optimizing DL models and facilitating the ongoing optimization of the DL models. © (2024), (Universitas Ahmad Dahlan). All rights reserved. |
publisher |
Universitas Ahmad Dahlan |
issn |
16936930 |
language |
English |
format |
Article |
accesstype |
|
record_format |
scopus |
collection |
Scopus |
_version_ |
1809678149440503808 |