HI-Audio | Hybrid and Interpretable Deep neural audio machines

Summary
Machine Listening, or AI for Sound, is defined as the general field of Artificial Intelligence applied to audio analysis, understanding and synthesis by a machine. The access to ever increasing super-computing facilities, combined with the availability of huge data repositories (although largely unannotated), has led to the emergence of a significant trend with pure data-driven machine learning approaches. The field has rapidly moved towards end-to-end neural approaches which aim to directly solve the machine learning problem for raw acoustic signals but often only loosely taking into account the nature and structure of the processed data. The main consequences are that the models are 1) overly complex, require massive amounts of data to be trained and extreme computing power to be efficient (in terms of task performance), and 2) remain largely unexplainable and non-interpretable. To overcome these major shortcomings, we believe that our prior knowledge about the nature of the processed data, their generation process and their perception by humans should be explicitly exploited in neural-based machine learning frameworks.
The aim of HI-Audio is to build such hybrid deep approaches combining parameter-efficient and interpretable signal models, musicological and physics-based models, with highly tailored, deep neural architectures. The research directions pursued in HI-Audio will exploit novel deterministic and statistical audio and sound environment models with dedicated neural auto-encoders and generative networks and target specific applications including speech and audio scene analysis, music information retrieval and sound transformation and synthesis.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: https://cordis.europa.eu/project/id/101052978
Start date: 01-10-2022
End date: 30-09-2027
Total budget - Public funding: 2 482 317,50 Euro - 2 482 317,00 Euro
Cordis data

Original description

Machine Listening, or AI for Sound, is defined as the general field of Artificial Intelligence applied to audio analysis, understanding and synthesis by a machine. The access to ever increasing super-computing facilities, combined with the availability of huge data repositories (although largely unannotated), has led to the emergence of a significant trend with pure data-driven machine learning approaches. The field has rapidly moved towards end-to-end neural approaches which aim to directly solve the machine learning problem for raw acoustic signals but often only loosely taking into account the nature and structure of the processed data. The main consequences are that the models are 1) overly complex, require massive amounts of data to be trained and extreme computing power to be efficient (in terms of task performance), and 2) remain largely unexplainable and non-interpretable. To overcome these major shortcomings, we believe that our prior knowledge about the nature of the processed data, their generation process and their perception by humans should be explicitly exploited in neural-based machine learning frameworks.
The aim of HI-Audio is to build such hybrid deep approaches combining parameter-efficient and interpretable signal models, musicological and physics-based models, with highly tailored, deep neural architectures. The research directions pursued in HI-Audio will exploit novel deterministic and statistical audio and sound environment models with dedicated neural auto-encoders and generative networks and target specific applications including speech and audio scene analysis, music information retrieval and sound transformation and synthesis.

Status

SIGNED

Call topic

ERC-2021-ADG

Update Date

09-02-2023
Images
No images available.
Geographical location(s)
Structured mapping
Unfold all
/
Fold all
Horizon Europe
HORIZON.1 Excellent Science
HORIZON.1.1 European Research Council (ERC)
HORIZON.1.1.0 Cross-cutting call topics
ERC-2021-ADG ERC ADVANCED GRANTS
HORIZON.1.1.1 Frontier science
ERC-2021-ADG ERC ADVANCED GRANTS