kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A novel autonomous learning framework to enhance sEMG-based hand gesture recognition using depth information
KTH, School of Engineering Sciences (SCI), Engineering Mechanics, Fluid Mechanics and Engineering Acoustics, Biomechanics. (MoveAbility Laboratory)ORCID iD: 0000-0001-8785-5885
Show others and affiliations
2021 (English)In: Biomedical Signal Processing and Control, ISSN 1746-8094, E-ISSN 1746-8108, Vol. 66, article id 102444Article in journal (Refereed) Published
Abstract [en]

Hand gesture recognition using surface electromyography (sEMG) has been one of the most efficient motion analysis techniques in human–computer interaction in the last few decades. In particular, multichannel sEMG techniques have achieved stable performance in hand gesture recognition. However, the general solution of collecting and labeling large data manually leads to time-consuming implementation. A novel learning method is therefore needed to facilitate efficient data collection and preprocessing. In this paper, a novel autonomous learning framework is proposed to integrate the benefits of both depth vision and EMG signals, which automatically label the class of collected EMG data using depth information. It then utilizes a multiple layer neural network (MNN) classifier to achieve real-time recognition of the hand gestures using only the sEMG. The overall framework is demonstrated in an augmented reality application by the recognition of 10 hand gestures using the Myo armband and an HTC VIVE PRO. The results show prominent performance by introducing depth information for real-time data labeling. 

Place, publisher, year, edition, pages
Elsevier BV , 2021. Vol. 66, article id 102444
Keywords [en]
Classification, Clustering, Depth vision, Hands gesture recognition, Machine learning, Augmented reality, Human computer interaction, Learning systems, Multilayer neural networks, Network layers, Palmprint recognition, Augmented reality applications, Autonomous learning, Computer interaction, Hand-gesture recognition, Motion analysis techniques, Real time recognition, Stable performance, Surface electromyography, Gesture recognition, article, classifier, electromyography, gesture, human, vision
National Category
Signal Processing Human Computer Interaction Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-305489DOI: 10.1016/j.bspc.2021.102444ISI: 000636240200045Scopus ID: 2-s2.0-85100728034OAI: oai:DiVA.org:kth-305489DiVA, id: diva2:1615455
Note

QC 20211130

Available from: 2021-11-30 Created: 2021-11-30 Last updated: 2022-06-29Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Zhang, Longbin

Search in DiVA

By author/editor
Zhang, Longbin
By organisation
Biomechanics
In the same journal
Biomedical Signal Processing and Control
Signal ProcessingHuman Computer InteractionComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 141 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf