3DInAction | Understanding human action from unstructured 3D point clouds using deep learning methods

Summary
Human action recognition and forecasting is an integral part of autonomous robotic systems that require human-robot interaction as well as other engineering problems. Action recognition is typically achieved using video data and deep learning methods. However, other tasks, e.g. classification, showed that it is often beneficial to additionally use 3D data. Namely, 3D point clouds that are sampled on the surfaces of objects and agents in the scene. Unfortunately, existing human action recognition methods are somewhat limited, motivating the following research. In this action, we describe a new class of algorithms for 3D human action recognition and forecasting using a deep learning-based approach. Our approach is novel in that it extends a recent body of work on action recognition from 2D to the 3D domain which is particularly challenging due to the unstructured, unordered and permutation invariant nature of 3D point clouds. Our algorithms use the global and local statistical properties of 3D point clouds along with a 3D convolutional neural network to devise novel multi-modal representation of human action. It is inherently robust to spatial changes in the 3D domain, unlike previous works which rely on the 2D projections. In practice, deep learning methods allow us to learn an inference model from real-world examples. A common methodology for action recognition includes creating an annotated dataset, training an inference model and testing its generalization. Our research objectives cover all of these tasks and suggest novel methods to tackle them. Overall, the proposed research offers a new point of view for these long-standing problems, and with the vast related work in other domains, it may bridge the gap to arrive at a generalizable, effective and efficient 3D human action recognition and forecasting machinery. The resulting algorithms may be used in several scientific and engineering domains such as human-robot interaction among other applications.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: https://cordis.europa.eu/project/id/893465
Start date: 01-01-2021
End date: 31-12-2023
Total budget - Public funding: 276 205,44 Euro - 276 205,00 Euro
Cordis data

Original description

Human action recognition and forecasting is an integral part of autonomous robotic systems that require human-robot interaction as well as other engineering problems. Action recognition is typically achieved using video data and deep learning methods. However, other tasks, e.g. classification, showed that it is often beneficial to additionally use 3D data. Namely, 3D point clouds that are sampled on the surfaces of objects and agents in the scene. Unfortunately, existing human action recognition methods are somewhat limited, motivating the following research. In this action, we describe a new class of algorithms for 3D human action recognition and forecasting using a deep learning-based approach. Our approach is novel in that it extends a recent body of work on action recognition from 2D to the 3D domain which is particularly challenging due to the unstructured, unordered and permutation invariant nature of 3D point clouds. Our algorithms use the global and local statistical properties of 3D point clouds along with a 3D convolutional neural network to devise novel multi-modal representation of human action. It is inherently robust to spatial changes in the 3D domain, unlike previous works which rely on the 2D projections. In practice, deep learning methods allow us to learn an inference model from real-world examples. A common methodology for action recognition includes creating an annotated dataset, training an inference model and testing its generalization. Our research objectives cover all of these tasks and suggest novel methods to tackle them. Overall, the proposed research offers a new point of view for these long-standing problems, and with the vast related work in other domains, it may bridge the gap to arrive at a generalizable, effective and efficient 3D human action recognition and forecasting machinery. The resulting algorithms may be used in several scientific and engineering domains such as human-robot interaction among other applications.

Status

CLOSED

Call topic

MSCA-IF-2019

Update Date

28-04-2024
Images
No images available.
Geographical location(s)
Structured mapping
Unfold all
/
Fold all
Horizon 2020
H2020-EU.1. EXCELLENT SCIENCE
H2020-EU.1.3. EXCELLENT SCIENCE - Marie Skłodowska-Curie Actions (MSCA)
H2020-EU.1.3.2. Nurturing excellence by means of cross-border and cross-sector mobility
H2020-MSCA-IF-2019
MSCA-IF-2019