First, please carefully read the call for artifacts and check if the artifact under evaluation adheres to the instructions for authors.
Below are the instructions for artifact reviewers.
There will be two rounds of reviews: the first round for mandatory artifacts of Regular Tool Papers, and the second round for voluntary artifacts of other papers. Please stay in contact during both AE periods as some reviewers may be assigned artifacts from both categories.
February 27, 2025 | Artifact submission deadline |
March 2, 2025 | Artifact bidding due |
March 6, 2025 | Smoke-test reviews due |
March 9, 2025 | Author responses due (with possible fixes to artifacts) |
March 19, 2025 | Full reviews due |
March 20-23, 2025 | AEC discussion |
March 24, 2025 | Artifact notification |
April 7, 2025 | Artifact submission deadline |
April 10, 2025 | Artifact bidding due |
April 14, 2025 | Smoke-test reviews due |
April 17, 2025 | Author responses due (with possible fixes to artifacts) |
April 27, 2025 | Full reviews due |
April 28-30, 2025 | AEC discussion |
May 1, 2025 | Artifact notification |
In artifact evaluation, we aim to evaluate the degree to which the artifact supports the claims of the paper but not the novelty or significance of the claims themselves. Therefore, you do not need to be an expert in the topics of the paper (as opposed to paper bidding). When placing your bids, please prioritize your ability and confidence to execute (and spot potential issues of) the artifact based on the its packaging (VM or container) and the environment (Linux, MacOS, or Windows) that the authors tested the artifact. In principle, please bid all artifacts that you are able/have enough resources to test.
Smoke tests are used to spot technical issues (e.g., installation) and allow authors to fix minor problems in their artifacts. The review should be uploaded on EasyChair, clearly indicating that it is related to the smoke test. In your smoke-test reviews, please report whether you are able to set up the artifact and run the tests provided by authors. Ideally, the authors should also provide the expected outputs of these tests so that reviewers can easily compare the obtained results on their machines to the expected results. If the smoke tests failed on your machines, clearly document the error messages for authors to debug. The smoke-test phase is the only interaction with the authors. Please make sure you can evaluate the artifact by the end of this phase.
In this stage, your goal is to reproduce the data reported in the paper by the artifact. The review should be uploaded on EasyChair (by editing the previous one), it is related to the full evaluation. Please read the evaluation section of the paper carefully to understand its experimental design. (It would be nice to also browse through the proposed approach, but it is not strictly necessary to figure every technical detail.) Identify the claims the artifact supports, and generate the data using the corresponding commands given by the authors. The authors should clearly state how to reproduce each claim in the paper, including any necessary post-processing steps. Remember to allocate enough time for the evaluation. Ideally, you should be able to reproduce the data and compare them to those reported by the authors.
In your final reviews, please clearly state which badges you recommend and the corresponding justifications. Please familiarize yourself with the EAPLS guidelines. Below we summarize the crucial assessment criteria:
Note on the consistency of results: It is very likely that the results reproduced by you with the artifact do not match exactly with the results reported in the paper (e.g., due to different machines). Therefore, when evaluating the consistency, focus on trends rather than the exact numbers.