Multi objective hyperparameter tuning via random search on deep learning models

This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and Ale...

Full description

Bibliographic Details
Published in:Telkomnika (Telecommunication Computing Electronics and Control)
Main Author: Rom A.R.M.; Jamil N.; Ibrahim S.
Format: Article
Language:English
Published: Universitas Ahmad Dahlan 2024
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde
id 2-s2.0-85197106107
spelling 2-s2.0-85197106107
Rom A.R.M.; Jamil N.; Ibrahim S.
Multi objective hyperparameter tuning via random search on deep learning models
2024
Telkomnika (Telecommunication Computing Electronics and Control)
22
4
10.12928/TELKOMNIKA.v22i4.25847
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde
This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and AlexNet implemented on prominent benchmark datasets of Modified National Institute of Standards and Technology (MNIST) and Canadian Institute for Advanced Research-10 (CIFAR-10). In the context of this study, the evaluation will be adopting a multi-objective framework, navigating the delicate trade-offs between conflicting performance metrics, including accuracy, F1-score, and model parameter size. The primary objective of employing a multi-objective evaluation framework is to enhance the understanding regarding the interactions of these performance metrics interact and influence each other. In real-world scenarios, DL models often need to strike a balance between these conflicting objectives. This research adds to the increasing wealth of knowledge in hyperparameter tuning for DL models and serves as a reference point for practitioners seeking to optimize their DL architectures. The results of our analysis are positioned to provide invaluable insights into the intricate balancing act required during the process of hyperparameter fine-tuning. These insights will contribute to the ongoing advancement of best practices in optimizing DL models and facilitating the ongoing optimization of the DL models. © (2024), (Universitas Ahmad Dahlan). All rights reserved.
Universitas Ahmad Dahlan
16936930
English
Article

author Rom A.R.M.; Jamil N.; Ibrahim S.
spellingShingle Rom A.R.M.; Jamil N.; Ibrahim S.
Multi objective hyperparameter tuning via random search on deep learning models
author_facet Rom A.R.M.; Jamil N.; Ibrahim S.
author_sort Rom A.R.M.; Jamil N.; Ibrahim S.
title Multi objective hyperparameter tuning via random search on deep learning models
title_short Multi objective hyperparameter tuning via random search on deep learning models
title_full Multi objective hyperparameter tuning via random search on deep learning models
title_fullStr Multi objective hyperparameter tuning via random search on deep learning models
title_full_unstemmed Multi objective hyperparameter tuning via random search on deep learning models
title_sort Multi objective hyperparameter tuning via random search on deep learning models
publishDate 2024
container_title Telkomnika (Telecommunication Computing Electronics and Control)
container_volume 22
container_issue 4
doi_str_mv 10.12928/TELKOMNIKA.v22i4.25847
url https://www.scopus.com/inward/record.uri?eid=2-s2.0-85197106107&doi=10.12928%2fTELKOMNIKA.v22i4.25847&partnerID=40&md5=90bc795eab60ec217b80b99d8cbfdcde
description This research examines the efficacy of random search (RS) in hyperparameter tuning, comparing its performance to baseline methods namely manual search and grid search. Our analysis spans various deep learning (DL) architectures-multilayer perceptron (MLP), convolutional neural network (CNN), and AlexNet implemented on prominent benchmark datasets of Modified National Institute of Standards and Technology (MNIST) and Canadian Institute for Advanced Research-10 (CIFAR-10). In the context of this study, the evaluation will be adopting a multi-objective framework, navigating the delicate trade-offs between conflicting performance metrics, including accuracy, F1-score, and model parameter size. The primary objective of employing a multi-objective evaluation framework is to enhance the understanding regarding the interactions of these performance metrics interact and influence each other. In real-world scenarios, DL models often need to strike a balance between these conflicting objectives. This research adds to the increasing wealth of knowledge in hyperparameter tuning for DL models and serves as a reference point for practitioners seeking to optimize their DL architectures. The results of our analysis are positioned to provide invaluable insights into the intricate balancing act required during the process of hyperparameter fine-tuning. These insights will contribute to the ongoing advancement of best practices in optimizing DL models and facilitating the ongoing optimization of the DL models. © (2024), (Universitas Ahmad Dahlan). All rights reserved.
publisher Universitas Ahmad Dahlan
issn 16936930
language English
format Article
accesstype
record_format scopus
collection Scopus
_version_ 1809678149440503808