SSDM | Deployable Decision-Making: Embracing Semantics for Robotic Safety in Everyday Scenarios

Summary
Recent breakthroughs in machine learning have opened up opportunities for robots to build a semantic understanding of their operating environment and interact with humans in more natural ways. While machine learning has unlocked new potentials for robot autonomy, as robots venture into the real world, physical interactions with the surrounding environment pose additional challenges. One typical challenge in practical applications is providing safety guarantees in robot decision-making. Much of the safe robot decision-making literature today focuses on explicit safety constraints defined in the robot state and input space. However, in practical applications, robots are often required to infer semantics-grounded safe actions from perception input. While recent machine learning techniques are increasingly capable of distilling semantic information from perception, translating the semantic understanding to explicit safety constraints is non-trivial. In this proposed project, we aim to close the perception-action loop and develop mathematical foundations and algorithmic tools that enable robots to make intelligent and semantically safe decisions.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: https://cordis.europa.eu/project/id/101155035
Start date: 01-06-2024
End date: 31-05-2026
Total budget - Public funding: - 173 847,00 Euro
Cordis data

Original description

Recent breakthroughs in machine learning have opened up opportunities for robots to build a semantic understanding of their operating environment and interact with humans in more natural ways. While machine learning has unlocked new potentials for robot autonomy, as robots venture into the real world, physical interactions with the surrounding environment pose additional challenges. One typical challenge in practical applications is providing safety guarantees in robot decision-making. Much of the safe robot decision-making literature today focuses on explicit safety constraints defined in the robot state and input space. However, in practical applications, robots are often required to infer semantics-grounded safe actions from perception input. While recent machine learning techniques are increasingly capable of distilling semantic information from perception, translating the semantic understanding to explicit safety constraints is non-trivial. In this proposed project, we aim to close the perception-action loop and develop mathematical foundations and algorithmic tools that enable robots to make intelligent and semantically safe decisions.

Status

SIGNED

Call topic

HORIZON-MSCA-2023-PF-01-01

Update Date

22-11-2024
Images
No images available.
Geographical location(s)
Structured mapping
Unfold all
/
Fold all
Horizon Europe
HORIZON.1 Excellent Science
HORIZON.1.2 Marie Skłodowska-Curie Actions (MSCA)
HORIZON.1.2.0 Cross-cutting call topics
HORIZON-MSCA-2023-PF-01
HORIZON-MSCA-2023-PF-01-01 MSCA Postdoctoral Fellowships 2023