Enhancing Sign Language Classification through Reservoir Computing and Depth Camera Technology

Thongking, Witchuda and Premachandra, Chinthaka and Wiranata, Ardi and Maeda, Shingo (2024) Enhancing Sign Language Classification through Reservoir Computing and Depth Camera Technology. In: 2024 International Conference on Image Processing and Robotics (ICIPRoB), 09-10 March 2024, Colombo, Sri Lanka.

[thumbnail of Enhancing_Sign_Language_Classification_through_Reservoir_Computing_and_Depth_Camera_Technology.pdf] Text
Enhancing_Sign_Language_Classification_through_Reservoir_Computing_and_Depth_Camera_Technology.pdf - Published Version
Restricted to Registered users only

Download (1MB) | Request a copy

Abstract

The depth camera has emerged as an efficacious tool for facilitating interactions between individuals and machines. Its advantageous features, including a lightweight design, high durability, completeness, superior image quality, and versatility, have prompted researchers to explore its integration into advanced applications. Currently, the prevailing methods for classifying sign language exhibit certain limitations, such as suboptimal accuracy and inefficiency in gesture recognition technologies. In this study, we conduct a comprehensive examination of hand language characteristics rooted in American Sign Language (ASL), utilizing gesture-based interactions and augmenting sensing performance through preprocessing and the application of reservoir computing in the training model. We conduct a comprehensive evaluation of the gesture characteristics captured by the depth camera, elucidating a methodology encompassing preprocessing, machine learning, and the technical prediction of salient gesture features derived from conventional depth camera video footage. This endeavor aims to establish a robust signal model for enhanced understanding and representation of the analyzed gestures. Our machine-learning models adeptly predict American Sign Language signs. Correlation values derived from the depth camera data demonstrate a pronounced alignment with naturally occurring variations in metrics observed within specific gestures. Our methodologies for quantifying gestures with depth cameras contribute to heightened accessibility in quantitative motion analysis. The outcomes of the reservoir computing implementation exhibit the successful classification of three hand language signs, thereby manifesting high precision and recall. The robust performance of the classification is pivotal for practical applications and underscores the efficacy of our approach. © 2024 IEEE.

Item Type: Conference or Workshop Item (Paper)
Additional Information: Cited by: 0
Uncontrolled Keywords: Cameras; Advanced applications; American sign language; Camera technology; Depth camera; High durability; Language sign; Lightweight design; Machine-learning; Reservoir Computing; Sign language; Machine learning
Subjects: T Technology > TJ Mechanical engineering and machinery
Divisions: Faculty of Engineering > Mechanical and Industrial Engineering Department
Depositing User: Rita Yulianti Yulianti
Date Deposited: 18 Feb 2025 04:28
Last Modified: 18 Feb 2025 04:28
URI: https://ir.lib.ugm.ac.id/id/eprint/13688

Actions (login required)

View Item
View Item