Explore the words cloud of the BigFastData project. It provides you a very rough idea of what is the project "BigFastData" about.
The following table provides information about the project.
Coordinator |
ECOLE POLYTECHNIQUE
Organization address contact info |
Coordinator Country | France [FR] |
Total cost | 2˙472˙752 € |
EC max contribution | 2˙472˙752 € (100%) |
Programme |
1. H2020-EU.1.1. (EXCELLENT SCIENCE - European Research Council (ERC)) |
Code Call | ERC-2016-COG |
Funding Scheme | ERC-COG |
Starting year | 2017 |
Duration (year-month-day) | from 2017-09-01 to 2022-08-31 |
Take a look of project's partnership.
# | ||||
---|---|---|---|---|
1 | ECOLE POLYTECHNIQUE | FR (PALAISEAU CEDEX) | coordinator | 2˙472˙752.00 |
This proposal addresses a pressing need from emerging big data applications such as genomics and data center monitoring: besides the scale of processing, big data systems must also enable perpetual, low-latency processing for a broad set of analytical tasks, referred to as big and fast data analysis. Today’s technology falls severely short for such needs due to the lack of support of complex analytics with scale, low latency, and strong guarantees of user performance requirements. To bridge the gap, this proposal tackles a grand challenge: “How do we design an algorithmic foundation that enables the development of all necessary pillars of big and fast data analysis?” This proposal considers three pillars: 1) Parallelism: There is a fundamental tension between data parallelism (for scale) and pipeline parallelism (for low latency). We propose new approaches based on intelligent use of memory and workload properties to integrate both forms of parallelism. 2) Analytics: The literature lacks a large body of algorithms for critical order-related analytics to be run under data and pipeline parallelism. We propose new algorithmic frameworks to enable such analytics. 3) Optimization: To run analytics, today's big data systems are best effort only. We transform such systems into a principled optimization framework that suits the new characteristics of big data infrastructure and adapts to meet user performance requirements.
The scale and complexity of the proposed algorithm design makes this project high-risk, at the same time, high-gain: it will lay a solid foundation for big and fast data analysis, enabling a new integrated parallel processing paradigm, algorithms for critical order-related analytics, and a principled optimizer with strong performance guarantees. It will also broadly enable accelerated information discovery in emerging domains such as genomics, as well as economic benefits of early, well-informed decisions and reduced user payments.
Are you the coordinator (or a participant) of this project? Plaese send me more information about the "BIGFASTDATA" project.
For instance: the website url (it has not provided by EU-opendata yet), the logo, a more detailed description of the project (in plain text as a rtf file or a word file), some pictures (as picture files, not embedded into any word file), twitter account, linkedin page, etc.
Send me an email (fabio@fabiodisconzi.com) and I put them in your project's page as son as possible.
Thanks. And then put a link of this page into your project's website.
The information about "BIGFASTDATA" are provided by the European Opendata Portal: CORDIS opendata.