- Sponsor:
- sighpc
WORKS'13 was the eighth workshop in the WORKS series. The call for papers attracted sixteen submissions from Asia, Europe, North and South America. The quality of the papers, peer reviewed by the program committee, were exceptional this year and overall thirteen of the papers were accepted, covering a variety of topics, including provenance and metadata, workflow models, workflow tasks characterization, workflow engines scalability, and distributed computing performance.
Proceeding Downloads
On assisting scientific data curation in collection-based dataflows using labels
Thanks to the proliferation of computational techniques and the availability of datasets, data-intensive research has become commonplace in science. Sharing and re-use of datasets is key to scientific progress. A critical requirement for enabling data ...
Static compiler analysis for workflow provenance
Data provenance is the lineage of an artifact or object. Provenance can provide a basis upon which data can be regenerated, and can be used to determine the quality of both the process and provenance itself. Provenance capture from workflows is ...
On specifying and sharing scientific workflow optimization results using research objects
- Sonja Holl,
- Daniel Garijo,
- Khalid Belhajjame,
- Olav Zimmermann,
- Renato De Giovanni,
- Matthias Obst,
- Carole Goble
Reusing and repurposing scientific workflows for novel scientific experiments is nowadays facilitated by workflow repositories. Such repositories allow scientists to find existing workflows and re-execute them. However, workflow input parameters often ...
Semantics and provenance for processing element composition in dispel workflows
Dispel is a scripting language for constructing workflow graph which can then be executed by some other computational infrastructure. It facilitates construction of abstract components (called Processing Elements, or PEs) that can be instantiated in ...
Execution time prediction for grid infrastructures based on runtime provenance data
An accurate performance prediction service can be very useful for resource management and the scheduler service and help them make better resource utilization decisions by providing better execution time estimates. In this paper we present a novel ...
Toward fine-grained online task characteristics estimation in scientific workflows
- Rafael Ferreira da Silva,
- Gideon Juve,
- Ewa Deelman,
- Tristan Glatard,
- Frédéric Desprez,
- Douglas Thain,
- Benjamin Tovar,
- Miron Livny
Task characteristics estimations such as runtime, disk space, and memory consumption, are commonly used by scheduling algorithms and resource provisioning techniques to provide successful and efficient workflow executions. These methods assume that ...
Understanding workflows for distributed computing: nitty-gritty details
- Silvia D. Olabarriaga,
- Mohammad Mahdi Jaghoori,
- Vladimir Korkhov,
- Barbera van Schaik,
- Antoine van Kampen
Scientific workflow management is heavily used in our organization. After six years, a large number of workflows are available and regularly used to run biomedical data analysis experiments on distributed infrastructures, mostly on grids. In this paper ...
A framework for dynamically generating predictive models of workflow execution
The ability to accurately predict the performance of software components executing within a Cloud environment is an area of intense interest to many researchers. The availability of an accurate prediction of the time taken for a piece of code to execute ...
Time-bound analytic tasks on large datasets through dynamic configuration of workflows
- Yolanda Gil,
- Varun Ratnakar,
- Rishi Verma,
- Andrew Hart,
- Paul Ramirez,
- Chris Mattmann,
- Arni Sumarlidason,
- Samuel L. Park
Domain experts are often untrained in big data technologies and this limits their ability to exploit the data they have available. Workflow systems hide the complexities of high-end computing and software engineering by offering pre-packaged analytic ...
Automated packaging of bioinformatics workflows for portability and durability using makeflow
Dependency management remains a major challenge for all forms of software. A program implemented in a given environment typically has many implicit dependencies on programs, libraries, and other objects present within that environment. Moving ...
Distributed tools deployment and management for multiple galaxy instances in globus genomics
- Dinanath Sulakhe,
- Alex Rodriguez,
- Nick Prozorovsky,
- Nilesh Kavthekar,
- Ravi Madduri,
- Amol Parikh,
- Paul Dave,
- Lukasz Lacinski,
- Ian Foster
Workflow systems play an important role in the analysis of the fast-growing genomics data produced by low-cost next generation sequencing (NGS) technologies. Many biomedical research groups lack the expertise to assemble and run the sophisticated ...
The demand for consistent web-based workflow editors
- Sandra Gesing,
- Malcolm Atkinson,
- Iraklis Klampanos,
- Michelle Galea,
- Michael R. Berthold,
- Roberto Barbera,
- Diego Scardaci,
- Gabor Terstyanszky,
- Tamas Kiss,
- Peter Kacsuk
This paper identifies the high value to researchers in many disciplines of having web-based graphical editors for scientific workflows and draws attention to two technological transitions: good quality editors can now run in a browser and workflow ...
Scalable script-based data analysis workflows on clouds
Data analysis workflows are often composed by many concurrent and compute-intensive tasks that can be efficiently executed only on scalable computing infrastructures, such as HPC systems, Grids and Cloud platforms. The use of Cloud services for the ...