Thongking, Witchuda and Premachandra, Chinthaka and Wiranata, Ardi and Maeda, Shingo (2024) Enhancing Sign Language Classification through Reservoir Computing and Depth Camera Technology. In: 2024 International Conference on Image Processing and Robotics (ICIPRoB), 09-10 March 2024, Colombo, Sri Lanka.
![[thumbnail of Enhancing_Sign_Language_Classification_through_Reservoir_Computing_and_Depth_Camera_Technology.pdf]](https://ir.lib.ugm.ac.id/style/images/fileicons/text.png)
Enhancing_Sign_Language_Classification_through_Reservoir_Computing_and_Depth_Camera_Technology.pdf - Published Version
Restricted to Registered users only
Download (1MB) | Request a copy
Abstract
The depth camera has emerged as an efficacious tool for facilitating interactions between individuals and machines. Its advantageous features, including a lightweight design, high durability, completeness, superior image quality, and versatility, have prompted researchers to explore its integration into advanced applications. Currently, the prevailing methods for classifying sign language exhibit certain limitations, such as suboptimal accuracy and inefficiency in gesture recognition technologies. In this study, we conduct a comprehensive examination of hand language characteristics rooted in American Sign Language (ASL), utilizing gesture-based interactions and augmenting sensing performance through preprocessing and the application of reservoir computing in the training model. We conduct a comprehensive evaluation of the gesture characteristics captured by the depth camera, elucidating a methodology encompassing preprocessing, machine learning, and the technical prediction of salient gesture features derived from conventional depth camera video footage. This endeavor aims to establish a robust signal model for enhanced understanding and representation of the analyzed gestures. Our machine-learning models adeptly predict American Sign Language signs. Correlation values derived from the depth camera data demonstrate a pronounced alignment with naturally occurring variations in metrics observed within specific gestures. Our methodologies for quantifying gestures with depth cameras contribute to heightened accessibility in quantitative motion analysis. The outcomes of the reservoir computing implementation exhibit the successful classification of three hand language signs, thereby manifesting high precision and recall. The robust performance of the classification is pivotal for practical applications and underscores the efficacy of our approach. © 2024 IEEE.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Additional Information: | Cited by: 0 |
Uncontrolled Keywords: | Cameras; Advanced applications; American sign language; Camera technology; Depth camera; High durability; Language sign; Lightweight design; Machine-learning; Reservoir Computing; Sign language; Machine learning |
Subjects: | T Technology > TJ Mechanical engineering and machinery |
Divisions: | Faculty of Engineering > Mechanical and Industrial Engineering Department |
Depositing User: | Rita Yulianti Yulianti |
Date Deposited: | 18 Feb 2025 04:28 |
Last Modified: | 18 Feb 2025 04:28 |
URI: | https://ir.lib.ugm.ac.id/id/eprint/13688 |