QEST + FORMATS 2026
🇬🇧 Liverpool, United KingdomSeptember 2-4, 2026

Artifact Evaluation

Reproducibility of experimental results is crucial to foster an atmosphere of trustworthy, open, and reusable research. To improve and reward reproducibility, QEST+FORMATS 2026 includes a dedicated Artifact Evaluation (AE). An artifact is any additional material (software, data sets, machine-checkable proofs, etc.) that supports the claims made in the paper and, in the ideal case, makes them fully reproducible. In case of a tool, a typical artifact consists of the binary or source code of the tool, its documentation, the input files (e.g., models analyzed or input data) used for the tool evaluation in the paper, and a configuration file or document describing the parameters used to obtain the results.

Submission of an artifact is mandatory for tool papers, and optional – but encouraged – for research papers if it can support the results presented in the paper. Artifacts will be reviewed concurrently to the corresponding papers. The results of the artifact evaluation will be taken into consideration in the paper reviewing discussion. However, the primary goal of the artifact evaluation is to give positive feedback to authors as well as encourage and reward reproducible research.

Benefits for Authors: By providing an artifact supporting experimental claims, authors increase the confidence of readers in their contribution. Accepted papers with a successfully evaluated artifact will receive a badge to be included on the paper’s title page. Finally, artifacts that significantly exceed expectations may receive an Outstanding Artifact Award.

Important Dates

All dates are AoE

Review phases of the evaluation process are explained below.

Artifact Evaluation

Criteria

The goal of this initiative is to encourage research that is openly accessible and reproducible also in the future (time-proof). The AE Committee (AEC) will evaluate the artifact according to the following criteria, which are based on the EAPLS guidelines.

For example, artifacts that need to download third-party material using a network connection, e.g. to retrieve external dependencies not bundled in the artifact, will not be considered future-proof. In contrast, if dependencies are bundled in the artifact and installed with automated script, the artifact can be considered future-proof.

Badging

For tool papers, submission of an artifact is mandatory. Artifacts for tool papers are expected to obtain the available and functional badge as a condition for acceptance.

For research papers and case study papers, submission of an artifact is optional-but strongly encouraged.

Papers that undergo a successful evaluation by the criteria above will be allowed to place a badge on the title page for the camera-ready version of their article. Sample LaTeX code to place the badge will be provided.

Note that the paper must clearly indicate the exact artifact version used for the evaluation, i.e. the DOI used for the submission. The paper may additionally link to the most recent version and/or source code repositories.

Process

The artifact evaluation is single blind. This in particular means that the artifact does not need to be anonymized.

The evaluation consists of two phases: the quick-check phase (Phase I) and the full-review phase (Phase II), which proceed as follows.

Outstanding Artifact Award

AEC members will nominate artifacts that significantly exceed expectations for an Outstanding Artifact Award. The AEC chairs will consider these nominations, and might award them to no or multiple artifacts. Awardees will receive a certificate during the social event.

Artifact Submission

An artifact submission should have the same authors and title (with suffix “(AE)”) as the paper submission. An artifact submission consists of:

Submissions are made through EasyChair at this link: https://easychair.org/conferences/?conf=qestformats2026 (Track: QEST+FORMATS 2026 AE).

The Artifact Itself

In the spirit of reproducibility and future-proofness, some requirements are imposed on the actual artifact. In case any of these points cannot be implemented, e.g. due to the use of licensed software that cannot be distributed, please contact the AEC chairs as soon as possible to discuss specific arrangements.

Contents of the artifact must include:

  1. A README file, describing in clear and simple steps how to install and use the artifact, how to reproduce the results presented in the paper, and how the files/output produced in the artifact corresponds to tables and figures in the paper:
    • The README file should provide instructions to run a quick check, such as a toy example or test run, to easily check the setup in Phase I;
    • In case network access is required by the artifact, an explanation of when and why it is required should be provided.
    • In case large amounts of resources are needed to run the full evaluation, we strongly recommend providing a way to replicate a representative subset of the results such that results can be reproduced on various hardware platforms in reasonable time (within about 8 hours on a modern notebook). Do include the full set of experiments for those reviewers with sufficient hardware or time, alongside an explanation on how to reproduce individual results.
  2. A LICENSE file, which at the very least allows the AEC to download and execute the artifact.
  3. The concrete binaries for experimental reproduction of results in the paper, either as a Virtual Machine (VM) or a Docker image, containing everything that is needed to run the artifact:
    • For Virtual Machine: use VirtualBox and save the VM as an Open Virtual Appliance (OVA) file. You can for example use this AE VM as basis;
    • For Docker: include the complete image saved with docker save (potentially compressed with e.g. gzip).

The artifact must be provided as a single compressed file, e.g. an artifact.zip or .rar/tar.gz/tar.xz archive, that contains a complete Docker image generated via docker save or a VirtualBox OVA file as indicated in point 3.

The artifact must be made available through archive-quality storage (Zenodo, Figshare, etc.) that provides a citable DOI.

The artifact must not require financial cost to be evaluated, e.g. by running on a cloud service.

It is also highly recommended, but not strictly required:

In general, it should be as simple as possible for reviewers to conclude reproducibility.

Sources and Reusability

Authors are also encouraged to include all sources, dependencies, and instructions needed to modify and/or (re-)build the artifact (e.g. through tarballs and a Dockerfile). This may, of course, rely on network access. We recommend to strive to be as self-contained as possible. In particular, the artifact should contain as much of its dependencies as reasonably possible, and any downloads should only refer to stable sources (e.g. use a standard Debian docker image as base) and precise versions (e.g. a concrete commit / tag in a GitHub repository or docker image version instead of an unspecific “latest”). This maximizes the chances that the tool can still be modified and built several years from now.