Summary
Machine learning and data science are areas of tremendous progress over the last decade, leading to exciting research developments, and significant practical impact. Broadly, progress in this area has been enabled by the rapidly increasing availability of data, by better algorithms, and by large-scale platforms enabling efficient computation on immense datasets. While it is reasonable to expect that the first two trends will continue for the foreseeable future, the same cannot be said of the third trend, of continually increasing computational performance. Increasing computational demands place immense pressure on algorithms and systems to scale, while the performance limits of traditional computing paradigms are becoming increasingly apparent. Thus, the question of building algorithms and systems for scalable machine learning is extremely pressing. The project will take a decisive step to answer this challenge, developing new abstractions, algorithms and system support for scalable machine learning. In a nutshell, the line of approach is elastic coordination: allowing machine learning algorithms to approximate and/or randomize their synchronization and communication semantics, in a structured, controlled fashion, to achieve scalability. The project exploits the insight that many such algorithms are inherently stochastic, and hence robust to inconsistencies. My thesis is that elastic coordination can lead to significant, consistent performance improvements across a wide range of applications, while guaranteeing provably correct answers. ScaleML will apply elastic coordination to two specific relevant scenarios: scalability inside a single multi-threaded machine, and scalability across networks of machines.
Conceptually, the project’s impact is in providing a set of new design principles and algorithms for scalable computation. It will develop these insights into a set of tools and working examples for scalable distributed machine learning.
Conceptually, the project’s impact is in providing a set of new design principles and algorithms for scalable computation. It will develop these insights into a set of tools and working examples for scalable distributed machine learning.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: | https://cordis.europa.eu/project/id/805223 |
Start date: | 01-03-2019 |
End date: | 29-02-2024 |
Total budget - Public funding: | 1 494 121,00 Euro - 1 494 121,00 Euro |
Cordis data
Original description
Machine learning and data science are areas of tremendous progress over the last decade, leading to exciting research developments, and significant practical impact. Broadly, progress in this area has been enabled by the rapidly increasing availability of data, by better algorithms, and by large-scale platforms enabling efficient computation on immense datasets. While it is reasonable to expect that the first two trends will continue for the foreseeable future, the same cannot be said of the third trend, of continually increasing computational performance. Increasing computational demands place immense pressure on algorithms and systems to scale, while the performance limits of traditional computing paradigms are becoming increasingly apparent. Thus, the question of building algorithms and systems for scalable machine learning is extremely pressing. The project will take a decisive step to answer this challenge, developing new abstractions, algorithms and system support for scalable machine learning. In a nutshell, the line of approach is elastic coordination: allowing machine learning algorithms to approximate and/or randomize their synchronization and communication semantics, in a structured, controlled fashion, to achieve scalability. The project exploits the insight that many such algorithms are inherently stochastic, and hence robust to inconsistencies. My thesis is that elastic coordination can lead to significant, consistent performance improvements across a wide range of applications, while guaranteeing provably correct answers. ScaleML will apply elastic coordination to two specific relevant scenarios: scalability inside a single multi-threaded machine, and scalability across networks of machines.Conceptually, the project’s impact is in providing a set of new design principles and algorithms for scalable computation. It will develop these insights into a set of tools and working examples for scalable distributed machine learning.
Status
SIGNEDCall topic
ERC-2018-STGUpdate Date
27-04-2024
Images
No images available.
Geographical location(s)