SN0526 : Building Efficient Data Planner for Peta-scale Science
Author(s) | : | Michal Zerola, Jérôme Lauret, Roman Barták, Michal Šumbera |
Date | : | Feb. 25, 2010 |
File(s) | : | zerola_acat2010.v5.pdf |
Abstract | : | Unprecedented data challenges both in terms of Peta-scale volume and concurrent distributed computing have seen birth with the rise of statistically driven experiments such as the ones represented by the high-energy and nuclear physics community. Distributed computing strategies, heavily relying on the presence of data at the proper place and time, have further raised demands for coordination of data movement on the road towards achieving high performance. Massive data processing will be hardly “fair” to users or unlikely be using network bandwidth efficiently whenever diverse usage patterns and priorities will be involved unless we address and deal with planning and reasoning of data movement and placement. Although there exist several sophisticated and efficient point-to-point data transfer tools, the lack of global planners and decision makers, answering questions such as “How to bring the required dataset to the user?” or “From which sources to grab the replicated data”, is for most part lacking. Submitted: ACAT 2010 proceedings |
Keywords | : | Scheduling, Planning, Mixed Integer Programming, neywork, data transfer |
Category | : | Computing |
- Login or register to post comments
- Back to STAR Notes page