Difference between revisions of "ExecutablePapers"
m |
|||
Line 1: | Line 1: | ||
== Introduction == | |||
While computational experiments have become an integral part of the scientific method, it is still a challenge to repeat such experiments, because often, computational experiments require specific hardware, non-trivial software installation, and complex manipulations to obtain results. We posit that integrating data acquisition, derivation, analysis, and visualization as executable components throughout the publication process will make it easier to generate and share repeatable results. | While computational experiments have become an integral part of the scientific method, it is still a challenge to repeat such experiments, because often, computational experiments require specific hardware, non-trivial software installation, and complex manipulations to obtain results. We posit that integrating data acquisition, derivation, analysis, and visualization as executable components throughout the publication process will make it easier to generate and share repeatable results. | ||
We have built an extensible infrastructure to support the life-cycle of ''executable publications''---their creation, review and re-use. | We have built an extensible infrastructure to support the life-cycle of ''executable publications''---their creation, review and re-use. | ||
Our focus is on papers whose ''computational experiments'' can be reproduced and validated. We note that our approach is orthogonal to others which focus on semantics and authoring, and can be combined with these. | Our focus is on papers whose ''computational experiments'' can be reproduced and validated. We note that our approach is orthogonal to others which focus on semantics and authoring, and can be combined with these. | ||
We have written a paper that details the challenges of computational repeatability and the solutions we have developed: | |||
http://www.cs.utah.edu/~juliana/pub/vistrails-executable-paper.pdf | http://www.cs.utah.edu/~juliana/pub/vistrails-executable-paper.pdf | ||
== Infrastructure == | |||
We have developed a set of techniques to help authors, reviewers, and readers construct and interact with executable papers. Many of these techniques have been integrated with the [http://www.vistrails.org/ VisTrails] system and the [http://crowdlabs.org crowdLabs] site. VisTrails is an open-source scientific workflow and provenance management system, and our extensions for executable papers allow users to create papers and Web publications whose figures and results are directly tied to the computations that generated them. Both the [http://www.latex-project.org LaTeX] and [http://www.mediawiki.org MediaWiki] extensions call VisTrails (either a server or local executable) to embed results into their documents. | |||
== Demostrations == | |||
* | To see our infrastructure in action, please see the following videos: | ||
* {{qt|link=http://www.vistrails.org/images/executable_paper_latex.mov|text=Editing an executable paper written using LaTeX and VisTrails}} | |||
* {{qt|link=http://www.vistrails.org/images/executable_paper_server.mov|text=Exploring a Web-hosted paper using server-based computation}} | |||
== Examples of Executable Publications == | == Examples of Executable Publications == |
Revision as of 16:46, 20 May 2011
Introduction
While computational experiments have become an integral part of the scientific method, it is still a challenge to repeat such experiments, because often, computational experiments require specific hardware, non-trivial software installation, and complex manipulations to obtain results. We posit that integrating data acquisition, derivation, analysis, and visualization as executable components throughout the publication process will make it easier to generate and share repeatable results. We have built an extensible infrastructure to support the life-cycle of executable publications---their creation, review and re-use. Our focus is on papers whose computational experiments can be reproduced and validated. We note that our approach is orthogonal to others which focus on semantics and authoring, and can be combined with these.
We have written a paper that details the challenges of computational repeatability and the solutions we have developed: http://www.cs.utah.edu/~juliana/pub/vistrails-executable-paper.pdf
Infrastructure
We have developed a set of techniques to help authors, reviewers, and readers construct and interact with executable papers. Many of these techniques have been integrated with the VisTrails system and the crowdLabs site. VisTrails is an open-source scientific workflow and provenance management system, and our extensions for executable papers allow users to create papers and Web publications whose figures and results are directly tied to the computations that generated them. Both the LaTeX and MediaWiki extensions call VisTrails (either a server or local executable) to embed results into their documents.
Demostrations
To see our infrastructure in action, please see the following videos:
Examples of Executable Publications
ALPS2.0
For a real example of an executable paper, whose results can be reproduced and validated, check out the PDF for ALPS2.0 paper at http://arxiv.org/pdf/1101.2646
To repeat the experiments shown in the paper, you will need to download VisTrails from: You can download VisTrails from: http://www.vistrails.org/index.php/Downloads The ALPS2.0 package is included in VisTrails 1.6.
CFD Flow Analysis
Another example of an executable publication, can be found at: http://www.vistrails.org/index.php/User:Tohline/CPM/Levels2and3 Because this paper is published on a Wiki, it is possible to interact with the results using a Web browser. Try it out!
WikiQuery
Here's a case study we did for the SIGMOD Repeatability effort: http://www.cs.utah.edu/~juliana/WikiQuery/WikiQuery_casestudy.pdf
For more details on how to run the WikiQuery experiments, and to obtain the experiments (data, code, workflows) see WikiQuery.