TY - GEN
T1 - Dilated Causal Convolution Based Human Activity Recognition Using Voxelized Point Cloud Radar Data
AU - Kakuba, Samuel
AU - Colaco, Savina Jassica
AU - Kim, Jung Hwan
AU - Lee, Dong Gyu
AU - Yoon, Young Jin
AU - Han, Dong Seog
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Due to the immense advantages that include contactless sensing, privacy-preserving, and lighting condition in-sensitivity, radar systems have been applied in Human Activity Recognition (HAR). The radar signal is often used in its raw form, pre-processed into micro-Doppler signatures or represented as voxelized Point clouds. However, the point cloud data is usually sparse and non-uniform. HAR deep learning models ought to learn the spatial and temporal features. These models should be robust for all considered activities and computationally efficient. Instead of other deep learning techniques used in literature, dilated causal convolutions (DCC) provide a broad receptive field with a few layers while preserving the resolution of the inputs throughout the model, thereby learning the spatial and temporal cues. In this paper, we investigated the use of DCC in combination with other deep learning techniques like residual blocks (RDCC), transformer encoders (TED), and bidirectional long-short-Term memory (BiLSTM). We subsequently proposed the DCCB model that consists of DCC layers and BiLSTM layers. The proposed model exhibits a commendable performance in terms of accuracy, and generalization especially in terms of balanced robustness for all activities.
AB - Due to the immense advantages that include contactless sensing, privacy-preserving, and lighting condition in-sensitivity, radar systems have been applied in Human Activity Recognition (HAR). The radar signal is often used in its raw form, pre-processed into micro-Doppler signatures or represented as voxelized Point clouds. However, the point cloud data is usually sparse and non-uniform. HAR deep learning models ought to learn the spatial and temporal features. These models should be robust for all considered activities and computationally efficient. Instead of other deep learning techniques used in literature, dilated causal convolutions (DCC) provide a broad receptive field with a few layers while preserving the resolution of the inputs throughout the model, thereby learning the spatial and temporal cues. In this paper, we investigated the use of DCC in combination with other deep learning techniques like residual blocks (RDCC), transformer encoders (TED), and bidirectional long-short-Term memory (BiLSTM). We subsequently proposed the DCCB model that consists of DCC layers and BiLSTM layers. The proposed model exhibits a commendable performance in terms of accuracy, and generalization especially in terms of balanced robustness for all activities.
KW - activity recognition
KW - dilated convolutions
KW - radar data
UR - http://www.scopus.com/inward/record.url?scp=85189939015&partnerID=8YFLogxK
U2 - 10.1109/ICAIIC60209.2024.10463502
DO - 10.1109/ICAIIC60209.2024.10463502
M3 - Conference contribution
AN - SCOPUS:85189939015
T3 - 6th International Conference on Artificial Intelligence in Information and Communication, ICAIIC 2024
SP - 812
EP - 815
BT - 6th International Conference on Artificial Intelligence in Information and Communication, ICAIIC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 6th International Conference on Artificial Intelligence in Information and Communication, ICAIIC 2024
Y2 - 19 February 2024 through 22 February 2024
ER -