Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN

In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the...

Full description

Bibliographic Details
Published in:2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024
Main Authors: Halim, Abdul Hadi Abdul; Ab Rahim, A'zraa Afhzan; Rozaini, Nur Athirah; Hassan, Siti Lailatul Mohd; Halim, Ili Shairah Abdul; Abdullah, Noor Ezan
Format: Proceedings Paper
Language:English
Published: IEEE 2024
Subjects:
Online Access:https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006
author Halim
Abdul Hadi Abdul; Ab Rahim
A'zraa Afhzan; Rozaini
Nur Athirah; Hassan
Siti Lailatul Mohd; Halim
Ili Shairah Abdul; Abdullah
Noor Ezan
spellingShingle Halim
Abdul Hadi Abdul; Ab Rahim
A'zraa Afhzan; Rozaini
Nur Athirah; Hassan
Siti Lailatul Mohd; Halim
Ili Shairah Abdul; Abdullah
Noor Ezan
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
Automation & Control Systems; Engineering
author_facet Halim
Abdul Hadi Abdul; Ab Rahim
A'zraa Afhzan; Rozaini
Nur Athirah; Hassan
Siti Lailatul Mohd; Halim
Ili Shairah Abdul; Abdullah
Noor Ezan
author_sort Halim
spelling Halim, Abdul Hadi Abdul; Ab Rahim, A'zraa Afhzan; Rozaini, Nur Athirah; Hassan, Siti Lailatul Mohd; Halim, Ili Shairah Abdul; Abdullah, Noor Ezan
Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024
English
Proceedings Paper
In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the study compares two algorithms: YOLOv5, known for its object detection capabilities, and the Convolutional Neural Network (CNN), which excels in learning features directly from raw images. The dataset, comprising ten common MSL phrases, was collected using a webcam and OpenCV from four non-deaf individuals, incorporating variations in hand gestures and lighting. The findings indicate that YOLOv5 significantly outperforms CNN, with YOLOv5 achieving an overall accuracy of 0.99, precision of 0.93, and recall of 0.99, while CNN achieves an overall accuracy of 0.94, precision of 0.76, and recall of 0.71. The study concludes that YOLOv5 offers a more robust and accurate model for MSL detection, which can serve as a valuable communication aid for MSL users. However, the study acknowledges limitations such as using a non-deaf contributor dataset and a modest sample size, which may not fully represent the diversity of MSL users. Future research is suggested to incorporate data from deaf individuals to enhance the model's applicability and accuracy, considering that signing patterns may vary between non-deaf and deaf signers.
IEEE
2638-1710

2024


10.1109/ICSGRC62081.2024.10691273
Automation & Control Systems; Engineering

WOS:001345150000006
https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006
title Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
title_short Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
title_full Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
title_fullStr Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
title_full_unstemmed Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
title_sort Malaysian Sign Language (MSL) Detection: Comparison of YOLOv5 and CNN
container_title 2024 IEEE 15TH CONTROL AND SYSTEM GRADUATE RESEARCH COLLOQUIUM, ICSGRC 2024
language English
format Proceedings Paper
description In addressing the underrepresentation of Malaysian Sign Language (MSL) in sign language detection research, this study evaluates the performance of deep learning models on an MSL dataset to achieve high accuracy, precision, and recall for hand gesture detection and translation. Methodologically, the study compares two algorithms: YOLOv5, known for its object detection capabilities, and the Convolutional Neural Network (CNN), which excels in learning features directly from raw images. The dataset, comprising ten common MSL phrases, was collected using a webcam and OpenCV from four non-deaf individuals, incorporating variations in hand gestures and lighting. The findings indicate that YOLOv5 significantly outperforms CNN, with YOLOv5 achieving an overall accuracy of 0.99, precision of 0.93, and recall of 0.99, while CNN achieves an overall accuracy of 0.94, precision of 0.76, and recall of 0.71. The study concludes that YOLOv5 offers a more robust and accurate model for MSL detection, which can serve as a valuable communication aid for MSL users. However, the study acknowledges limitations such as using a non-deaf contributor dataset and a modest sample size, which may not fully represent the diversity of MSL users. Future research is suggested to incorporate data from deaf individuals to enhance the model's applicability and accuracy, considering that signing patterns may vary between non-deaf and deaf signers.
publisher IEEE
issn 2638-1710

publishDate 2024
container_volume
container_issue
doi_str_mv 10.1109/ICSGRC62081.2024.10691273
topic Automation & Control Systems; Engineering
topic_facet Automation & Control Systems; Engineering
accesstype
id WOS:001345150000006
url https://www-webofscience-com.uitm.idm.oclc.org/wos/woscc/full-record/WOS:001345150000006
record_format wos
collection Web of Science (WoS)
_version_ 1823296085372698624