초기 융합 기반 다중 모드 딥 러닝을 사용한 손 제스처 분류

Translated title of the contribution: Hand gesture classification using early fusion based multimodal deep learning

Ik Jin Kim, Su Yeol Kim, Yong Chan Lee, Yun Jung Lee

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, we propose a new hand gesture classification strategy using early fusion based multimodal deep learning. The structure and parameters of the state-of-the-art deep learning models such as ResNet152, DenseNet201, EfficientNetB0 for the source task of image classification are reused in the target task of hand gesture classification using surface electromyograph(EMG) and finger's kinematic data. The time-domain EMG and kinematic signals are normalized and then transformed into combined 2-D images for the early-fusion network. The experimental results support the superiority of the proposed method in terms of classification accuracy. The transfer learning model with the EfficientNetB0 shows the 93.94% accuracy for 40 gestures of 40 participants in the Ninapro DB2.

Translated title of the contributionHand gesture classification using early fusion based multimodal deep learning
Original languageKorean
Pages (from-to)1714-1721
Number of pages8
JournalTransactions of the Korean Institute of Electrical Engineers
Volume70
Issue number11
DOIs
StatePublished - Nov 2021

Keywords

  • Deep Learning
  • EMG
  • Hand Gesture Classification
  • Multimodal Learning
  • Ninapro DB

Fingerprint

Dive into the research topics of 'Hand gesture classification using early fusion based multimodal deep learning'. Together they form a unique fingerprint.

Cite this