Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the...
Published in: | 2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024 |
---|---|
Main Authors: | , , , , , , |
Format: | Proceedings Paper |
Language: | English |
Published: |
IEEE
2024
|
Subjects: | |
Online Access: | https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006 |
author |
Halim Abdul Hadi Abdul; Ab Rahim A'zraa Afhzan; Rozaini Nur Athirah; Hassan Siti Lailatul Mohd; Halim Ili Shairah Abdul; Abdullah Noor Ezan |
---|---|
spellingShingle |
Halim Abdul Hadi Abdul; Ab Rahim A'zraa Afhzan; Rozaini Nur Athirah; Hassan Siti Lailatul Mohd; Halim Ili Shairah Abdul; Abdullah Noor Ezan Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN Automation & Control Systems; Engineering |
author_facet |
Halim Abdul Hadi Abdul; Ab Rahim A'zraa Afhzan; Rozaini Nur Athirah; Hassan Siti Lailatul Mohd; Halim Ili Shairah Abdul; Abdullah Noor Ezan |
author_sort |
Halim |
spelling |
Halim, Abdul Hadi Abdul; Ab Rahim, A'zraa Afhzan; Rozaini, Nur Athirah; Hassan, Siti Lailatul Mohd; Halim, Ili Shairah Abdul; Abdullah, Noor Ezan Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN 2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024 English Proceedings Paper In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the study compares two algorithms: YOLOv5, known for its object detection capabilities, and the Convolutional Neural Network (CNN), which excels in learning features directly from raw images. The dataset, comprising ten common MSL phrases, was collected using a webcam and OpenCV from four non-deaf individuals, incorporating variations in hand gestures and lighting. The findings indicate that YOLOv5 significantly outperforms CNN, with YOLOv5 achieving an overall accuracy of 0.99, precision of 0.93, and recall of 0.99, while CNN achieves an overall accuracy of 0.94, precision of 0.76, and recall of 0.71. The study concludes that YOLOv5 offers a more robust and accurate model for MSL detection, which can serve as a valuable communication aid for MSL users. However, the study acknowledges limitations such as using a non-deaf contributor dataset and a modest sample size, which may not fully represent the diversity of MSL users. Future research is suggested to incorporate data from deaf individuals to enhance the model's applicability and accuracy, considering that signing patterns may vary between non-deaf and deaf signers. IEEE 2638-1710 2024 10.1109/ICSGRC62081.2024.10691273 Automation & Control Systems; Engineering WOS:001345150000006 https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006 |
title |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
title_short |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
title_full |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
title_fullStr |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
title_full_unstemmed |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
title_sort |
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN |
container_title |
2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024 |
language |
English |
format |
Proceedings Paper |
description |
In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the study compares two algorithms: YOLOv5, known for its object detection capabilities, and the Convolutional Neural Network (CNN), which excels in learning features directly from raw images. The dataset, comprising ten common MSL phrases, was collected using a webcam and OpenCV from four non-deaf individuals, incorporating variations in hand gestures and lighting. The findings indicate that YOLOv5 significantly outperforms CNN, with YOLOv5 achieving an overall accuracy of 0.99, precision of 0.93, and recall of 0.99, while CNN achieves an overall accuracy of 0.94, precision of 0.76, and recall of 0.71. The study concludes that YOLOv5 offers a more robust and accurate model for MSL detection, which can serve as a valuable communication aid for MSL users. However, the study acknowledges limitations such as using a non-deaf contributor dataset and a modest sample size, which may not fully represent the diversity of MSL users. Future research is suggested to incorporate data from deaf individuals to enhance the model's applicability and accuracy, considering that signing patterns may vary between non-deaf and deaf signers. |
publisher |
IEEE |
issn |
2638-1710 |
publishDate |
2024 |
container_volume |
|
container_issue |
|
doi_str_mv |
10.1109/ICSGRC62081.2024.10691273 |
topic |
Automation & Control Systems; Engineering |
topic_facet |
Automation & Control Systems; Engineering |
accesstype |
|
id |
WOS:001345150000006 |
url |
https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006 |
record_format |
wos |
collection |
Web of Science (WoS) |
_version_ |
1823296085372698624 |