Summary
Machine learning is revolutionising computer science and AI. Much of its success is due to deep neural networks, which have demonstrated outstanding performance in perception tasks such as image classification. Solutions based on deep learning are now being deployed in real-world systems, from virtual personal assistants to self-driving cars. Unfortunately, the black-box nature and instability of deep neural networks is raising concerns about the readiness of this technology. Efforts to address robustness of deep learning are emerging, but are limited to simple properties and function-based perception tasks that learn data associations. While perception is an essential feature of an artificial agent, achieving beneficial collaboration between human and artificial agents requires models of autonomy, inference, decision making, control and coordination that significantly go beyond perception. To address this challenge, this project will capitalise on recent breakthroughs by the PI and develop a model-based, probabilistic reasoning framework for autonomous agents with cognitive aspects, which supports reasoning about their decisions, agent interactions and inferences that capture cognitive information, in presence of uncertainty and partial observability. The objectives are to develop novel probabilistic verification and synthesis techniques to guarantee safety, robustness and fairness for complex decisions based on machine learning, formulate a comprehensive, compositional game-based modelling framework for reasoning about systems of autonomous agents and their interactions, and evaluate the techniques on a variety of case studies.
Addressing these challenges will require a fundamental shift towards Bayesian methods, and development of new, scalable, techniques, which differ from conventional probabilistic verification. If successful, the project will result in major advances in the quest towards provably robust and beneficial AI.
Addressing these challenges will require a fundamental shift towards Bayesian methods, and development of new, scalable, techniques, which differ from conventional probabilistic verification. If successful, the project will result in major advances in the quest towards provably robust and beneficial AI.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: | https://cordis.europa.eu/project/id/834115 |
Start date: | 01-10-2019 |
End date: | 30-09-2025 |
Total budget - Public funding: | 2 417 891,50 Euro - 2 417 890,00 Euro |
Cordis data
Original description
Machine learning is revolutionising computer science and AI. Much of its success is due to deep neural networks, which have demonstrated outstanding performance in perception tasks such as image classification. Solutions based on deep learning are now being deployed in real-world systems, from virtual personal assistants to self-driving cars. Unfortunately, the black-box nature and instability of deep neural networks is raising concerns about the readiness of this technology. Efforts to address robustness of deep learning are emerging, but are limited to simple properties and function-based perception tasks that learn data associations. While perception is an essential feature of an artificial agent, achieving beneficial collaboration between human and artificial agents requires models of autonomy, inference, decision making, control and coordination that significantly go beyond perception. To address this challenge, this project will capitalise on recent breakthroughs by the PI and develop a model-based, probabilistic reasoning framework for autonomous agents with cognitive aspects, which supports reasoning about their decisions, agent interactions and inferences that capture cognitive information, in presence of uncertainty and partial observability. The objectives are to develop novel probabilistic verification and synthesis techniques to guarantee safety, robustness and fairness for complex decisions based on machine learning, formulate a comprehensive, compositional game-based modelling framework for reasoning about systems of autonomous agents and their interactions, and evaluate the techniques on a variety of case studies.Addressing these challenges will require a fundamental shift towards Bayesian methods, and development of new, scalable, techniques, which differ from conventional probabilistic verification. If successful, the project will result in major advances in the quest towards provably robust and beneficial AI.
Status
SIGNEDCall topic
ERC-2018-ADGUpdate Date
27-04-2024
Images
No images available.
Geographical location(s)