TY - GEN
T1 - ArSL21L
T2 - 13th IEEE Global Engineering Education Conference, EDUCON 2022
AU - Batnasan, Ganzorig
AU - Gochoo, Munkhjargal
AU - Otgonbold, Munkh Erdene
AU - Alnajjar, Fady
AU - Shih, Timothy K.
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - It is complicated for the PwHL (people with hearing loss) to make a relationship with social majority, which naturally demands an interactive auto computer systems that have ability to understand sign language. With a trending Metaverse applications using augmented reality (AR) and virtual reality (VR), it is easier and interesting to teach sign language remotely using an avatar that mimics the gesture of a person using AI (Artificial Intelligence)-based system. There are various proposed methods and datasets for English SL (sign language); however, it is limited for Arabic sign language. Therefore, we present our collected and annotated Arabic Sign Language Letters Dataset (ArSL21L) consisting of 14202 images of 32 letter signs with various backgrounds collected from 50 people. We benchmarked our ArSL21L dataset on state-of-the-art object detection models, i.e., 4 versions of YOLOv5. Among the models, YOLOv5l achieved the best result with COCOmAP of 0.83. Moreover, we provide comparison results of classification task between ArSL2018 dataset, the only Arabic sign language letter dataset for classification task, and our dataset by running classification task on in-house short video. The results revealed that the model trained on our dataset has a superior performance over the model trained on ArSL2018. Moreover, we have created our prototype avatar which can mimic the ArSL (Arabic Sign Language) gestures for Metaverse applications. Finally, we believe, ArSL21L and the ArSL avatar will offer an opportunity to enhance the research and educational applications for not only the PwHL, but also in general real and virtual world applications. Our ArSL21L benchmark dataset is publicly available for research use on the Mendeley.
AB - It is complicated for the PwHL (people with hearing loss) to make a relationship with social majority, which naturally demands an interactive auto computer systems that have ability to understand sign language. With a trending Metaverse applications using augmented reality (AR) and virtual reality (VR), it is easier and interesting to teach sign language remotely using an avatar that mimics the gesture of a person using AI (Artificial Intelligence)-based system. There are various proposed methods and datasets for English SL (sign language); however, it is limited for Arabic sign language. Therefore, we present our collected and annotated Arabic Sign Language Letters Dataset (ArSL21L) consisting of 14202 images of 32 letter signs with various backgrounds collected from 50 people. We benchmarked our ArSL21L dataset on state-of-the-art object detection models, i.e., 4 versions of YOLOv5. Among the models, YOLOv5l achieved the best result with COCOmAP of 0.83. Moreover, we provide comparison results of classification task between ArSL2018 dataset, the only Arabic sign language letter dataset for classification task, and our dataset by running classification task on in-house short video. The results revealed that the model trained on our dataset has a superior performance over the model trained on ArSL2018. Moreover, we have created our prototype avatar which can mimic the ArSL (Arabic Sign Language) gestures for Metaverse applications. Finally, we believe, ArSL21L and the ArSL avatar will offer an opportunity to enhance the research and educational applications for not only the PwHL, but also in general real and virtual world applications. Our ArSL21L benchmark dataset is publicly available for research use on the Mendeley.
KW - Arabic sign language
KW - YOLOv5
KW - benchmarking
KW - deep learning
KW - image classification
KW - metaverse
KW - object detection
UR - http://www.scopus.com/inward/record.url?scp=85130440819&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85130440819&partnerID=8YFLogxK
U2 - 10.1109/EDUCON52537.2022.9766497
DO - 10.1109/EDUCON52537.2022.9766497
M3 - Conference contribution
AN - SCOPUS:85130440819
T3 - IEEE Global Engineering Education Conference, EDUCON
SP - 1814
EP - 1821
BT - Proceedings of the 2022 IEEE Global Engineering Education Conference, EDUCON 2022
A2 - Jemni, Mohammed
A2 - Kallel, Ilhem
A2 - Akkari, Abdeljalil
PB - IEEE Computer Society
Y2 - 28 March 2022 through 31 March 2022
ER -