Decoding Electromyography Signal With Multiple Labels for Hand Gesture Recognition

Decoding Electromyography Signal With Multiple Labels for Hand Gesture RecognitionDecoding Electromyography Signal With Multiple Labels for Hand Gesture Recognition

Abstract:

Surface electromyography (sEMG) is a significant interaction signal in the fields of human-computer interaction and rehabilitation assessment, as it can be used for hand gesture recognition. This letter proposes a novel MLHG model to improve the robustness of sEMG-based hand gesture recognition. The model utilizes multiple labels to decode the sEMG signals from two different perspectives. In the first view, the sEMG signals are transformed into motion signals using the proposed FES-MSCNN (Feature Extraction of sEMG with Multiple Sub-CNN modules). Furthermore, a discriminator FEM-SAGE (Feature Extraction of Motion with graph SAmple and aggreGatE model) is employed to judge the authenticity of the generated motion data. The deep features of the motion signals are extracted using the FEM-SAGE model. In the second view, the deep features of the sEMG signals are extracted using the FES-MSCNN model. The extracted features of the sEMG signals and the generated motion signals are then fused for hand gesture recognition. To evaluate the performance of the proposed model, a dataset containing sEMG signals and multiple labels from 12 subjects has been collected. The experimental results indicate that the MLHG model achieves an accuracy of 99.26%99.26% for within-session hand gesture recognition, 78.47%78.47% for cross-time, and 53.52%53.52% for cross-subject. These results represent a significant improvement compared to using only the gesture labels, with accuracy improvements of 1.91%1.91% , 5.35%5.35% , and 5.25%5.25% in the within-session, cross-time and cross-subject cases, respectively.