arXiv Analytics

Sign in

arXiv:2307.09238 [cs.CV]AbstractReferencesReviewsResources

Fusing Hand and Body Skeletons for Human Action Recognition in Assembly

Dustin Aganian, Mona Köhler, Benedict Stephan, Markus Eisenbach, Horst-Michael Gross

Published 2023-07-18Version 1

As collaborative robots (cobots) continue to gain popularity in industrial manufacturing, effective human-robot collaboration becomes crucial. Cobots should be able to recognize human actions to assist with assembly tasks and act autonomously. To achieve this, skeleton-based approaches are often used due to their ability to generalize across various people and environments. Although body skeleton approaches are widely used for action recognition, they may not be accurate enough for assembly actions where the worker's fingers and hands play a significant role. To address this limitation, we propose a method in which less detailed body skeletons are combined with highly detailed hand skeletons. We investigate CNNs and transformers, the latter of which are particularly adept at extracting and combining important information from both skeleton types using attention. This paper demonstrates the effectiveness of our proposed approach in enhancing action recognition in assembly scenarios.

Comments: International Conference on Artificial Neural Networks (ICANN) 2023
Categories: cs.CV, cs.LG, cs.RO
Related articles: Most relevant | Search more
arXiv:2010.16073 [cs.CV] (Published 2020-10-29)
CNN based Multistage Gated Average Fusion (MGAF) for Human Action Recognition Using Depth and Inertial Sensors
arXiv:2105.13533 [cs.CV] (Published 2021-05-28)
Inertial Sensor Data To Image Encoding For Human Action Recognition
arXiv:2407.06162 [cs.CV] (Published 2024-06-02)
RNNs, CNNs and Transformers in Human Action Recognition: A Survey and A Hybrid Model