-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathartefacts_submission_reviewing_guidelines.tex
815 lines (690 loc) · 37.1 KB
/
artefacts_submission_reviewing_guidelines.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
\documentclass[twoside,a4paper]{refart}
\usepackage{fontspec}
\setromanfont{Roboto}
\setmonofont{Roboto Mono}
\usepackage{xspace}
\usepackage[dvipsnames]{xcolor}
\usepackage{fontawesome}
\usepackage{marvosym}
\usepackage{hyperref}
\hypersetup{
colorlinks=true,
linkcolor=blue,
filecolor=magenta,
urlcolor=cyan
}
\setcounter{tocdepth}{2}
\setcounter{secnumdepth}{1}
\usepackage{graphicx}
\usepackage{tabularx}
\usepackage{makeidx}
\usepackage{ifthen}
% ifthen wird vom Bild von N.Beebe gebraucht!
\def\bs{\char'134 } % backslash in \tt font.
\newcommand{\ie}{i.\,e.,\xspace}
\newcommand{\eg}{e.\,g.,\xspace}
\newcommand{\evaluated}{\protect\hyperlink{evaluated}{Artifacts Evaluated}\xspace}
\newcommand{\functional}{\protect\hyperlink{functional}{Functional}\xspace}
\newcommand{\functionalfull}{\protect\hyperlink{functional}{Artifacts Evaluated - Functional}\xspace}
\newcommand{\reusable}{\protect\hyperlink{reusable}{Reusable}\xspace}
\newcommand{\reusablefull}{\protect\hyperlink{reusable}{Artifacts Evaluated - Reusable}\xspace}
\newcommand{\available}{\protect\hyperlink{available}{Artifacts Available}\xspace}
\newcommand{\validated}{\protect\hyperlink{validated}{Results Validated}\xspace}
\newcommand{\reproduced}{\protect\hyperlink{reproduced}{Results Reproduced}\xspace}
\newcommand{\replicated}{\protect\hyperlink{replicated}{Results Replicated}\xspace}
\newcommand{\documented}{\protect\hyperlink{documented}{documented}\xspace}
\newcommand{\consistent}{\protect\hyperlink{consistent}{consistent}\xspace}
\newcommand{\complete}{\protect\hyperlink{complete}{complete}\xspace}
\newcommand{\exercisable}{\protect\hyperlink{exercisable}{exercisable}\xspace}
\newcommand{\ipackage}{\protect\hyperlink{ipackage}{Instalation Package}\xspace}
\newcommand{\ipackages}{\protect\hyperlink{ipackage}{Instalation Packages}\xspace}
\newcommand{\spackage}{\protect\hyperlink{spackage}{Simple Package}\xspace}
\newcommand{\spackages}{\protect\hyperlink{spackage}{Simple Packages}\xspace}
\newcommand{\deadline}[1]{\attention[\faCalendar]{#1}}
\DeclareRobustCommand\cs[1]{\texttt{\char`\\#1}}
\title{FSE 2021 Artifact Evaluation Track:\\
Submission and Reviewing Guidelines}
\author{Silvia Abrahão and Daniel Mendez, amended for FSE 2021 by
Panos Louridas}
\begin{document}
\maketitle
\subsection*{Scope and Objectives}
This document outlines, briefly, the submission and reviewing process
for the Artifact Evaluation track of FSE 2021. It aims at
providing authors and reviewers with pragmatic insights into the
process and expected criteria to merit awarding the respective badges.
\subsection*{General Remarks on the Artifact Evaluation Track and
Expected Attitude}
In principle, the goal of the track is to promote and celebrate open
science. We therefore understand the track as one important means to
actively engage with the community in order to support them in making
their research artifacts publicly available and in fostering
replication of research results. The final result of the artifact
evaluation is to reward (only) the authors' work that satisfies the
criteria listed below with a set of qualifying badges as a form of
recognition. Yet, we see the track and the review phase as a unique
chance to actively support the research community in open science, so
instead of reviewing the artifacts ``blindly'' according to the
evaluation criteria towards the end of the review phase and submitting
a review with a go/no-go decision, we encourage all reviewers to make
use of the rebuttal phase in order to actively support the authors in
improving their submissions, same as we encourage authors to actively
engage with the reviewers and do their best to address their
well-intended suggestions as efficiently as possible.
\newpage
\tableofcontents
\newpage
\section{Badges Overview and Eligibility}
The artifact evaluation track aims to review, promote, share, and
catalog the research artifacts of accepted software engineering
papers. Authors of an accepted \emph{research paper} can submit an
artifact for the \evaluated and \available badges. Authors of any
prior SE work (published at FSE or elsewhere) are also invited to
submit their work for the \validated badges. Definitions for the
badges are given in the table below, taken from the
\href{https://www.acm.org/publications/policies/artifact-review-and-badging-current}{ACM
Artifact Review and Badging Version~1.1}. The top two artifacts
selected by the Programme Committee will be awarded the best artifact
awards. All accepted abstracts documenting the artifacts will be
published in the FSE 2021 proceedings as a further form of
recognition.
\vspace{2em}
\hspace{-1.45\marginparwidth}
\begin{tabularx}{0.9\paperwidth}{XXXXX}
\multicolumn{2}{c}{\hypertarget{evaluated}{\evaluated}}
& \centering \raisebox{0pt}{\hypertarget{available}{\available}}
& \multicolumn{2}{c}{\hypertarget{validated}{\validated}}
\arraybackslash \\
\centering \hypertarget{functional}{\functional}
&
\centering \hypertarget{reusable}{\reusable}
&
& \centering \hypertarget{reproduced}{\reproduced}
& \centering \hypertarget{replicated}{\replicated} \arraybackslash
\\ [2em]
\centering
\includegraphics[width=2cm]{artifacts_evaluated_functional_v1_1.png}
&
\centering
\includegraphics[width=2cm]{artifacts_evaluated_reusable_v1_1.png}
&
\centering
\includegraphics[width=2cm]{artifacts_available_v1_1.png}
&
\centering
\includegraphics[width=2cm]{results_reproduced_v1_1.png}
&
\centering
\includegraphics[width=2cm]{results_replicated_v1_1.png}
\arraybackslash
\\[2em]
\raggedright The artifacts associated with the research are found to
be \documented, \consistent, \complete, \exercisable, and include
appropriate evidence of verification and validation.
& \raggedright \functional +\newline
the artifacts associated with the
paper are of a quality that significantly exceeds minimal
functionality. They are very carefully documented and
well-structured to the extent that reuse and repurposing is
facilitated. In particular, norms and standards of the research
community for artifacts of this type are strictly followed.
& \raggedright Author-created artifacts relevant to this paper
have been placed on a publicly accessible archival repository. A
DOI or link to this repository along with a unique identifier for
the object is provided.
& \raggedright Τhe main results of the paper have been obtained in
a subsequent study by a person or team other than the original
authors, using, in part, artifacts provided by the original
authors.
& \raggedright Τhe main results of the paper have been
independently obtained in a subsequent study by a person or team
other than the original authors, without the use of author-supplied
artifacts.
\end{tabularx}
\newpage
\subsection{Badges in More Detail}
There are three different badges, two of which distinguish between two
levels.
\marginlabel{\evaluated:} This badge is applied to papers whose
associated artifacts have successfully completed an independent audit.
Artifacts need not be made publicly available to be considered for
this badge. However, they do need to be made available to reviewers.
Two levels are distinguished, only one of which should be applied in
any instance:
\begin{itemize}
\item \functionalfull These artifacts need to be:
\begin{itemize}
\item \hypertarget{documented}{\documented}: At minimum, an inventory
of artifacts is included, and sufficient description provided to
enable the artifacts to be exercised.
\item \hypertarget{consistent}{\consistent}: The artifacts are
relevant to the associated paper, and contribute in some inherent
way to the generation of its main results.
\item \hypertarget{complete}{\complete}: To the extent possible, all
components relevant to the paper in question are included.
(Proprietary artifacts need not be included. If they are required
to exercise the package then this should be documented, along with
instructions on how to obtain them. Proxies for proprietary data
should be included so as to demonstrate the analysis.)
\item \hypertarget{exercisable}{\exercisable}: Included scripts and
/ or software used to generate the results in the associated paper
can be successfully executed, and included data can be accessed
and appropriately manipulated.
\end{itemize}
\item \reusablefull The artifacts meet the requirements for the
\functionalfull level and in addition they are of a quality that
significantly exceeds the requirements set for the first level.
\emph{Authors are strongly encouraged to target their artifact
submissions for \reusablefull}, as the purpose of artifact
badges is, among other things, to facilitate reuse and
repurposing, which may not be achieved at the \functionalfull
level.
\end{itemize}
\marginlabel{\available:} This badge is applied to papers in which
associated artifacts have been made permanently available for
retrieval.
\begin{itemize}
\item We consider temporary drives (\eg Dropbox, Google Drive) to be
non-persistent, same as individual/institutional websites of the
submitting authors, as these are prone to changes.
\item We do not mandate the use of specific repositories. Although not
limited to, we strongly recommend relying on services like Zenodo to
archiving repositories / repository releases (\eg from GitHub) as
these services are persistent and they also offer the possibility to
assign a DOI. In principle, however, publisher repositories (\eg
ACM Digital Library) and open commercial repositories (\eg
figshare) are acceptable as well as long as they offer a declared
plan to enable permanent accessibility.
\item Artifacts do not need to have been formally evaluated in order
for an article to receive this badge. In addition, they need not be
\complete in the sense described above. They simply need to be
relevant to the study and add value beyond the text in the article.
Such artifacts could be something as simple as the data from which
the figures are drawn, or as complex as a complete software system
under study.
\end{itemize}
\marginlabel{\validated:} This badge is applied to papers in which the
main results of the paper have been successfully obtained by a person
or team other than the author. Two levels are distinguished, only one
of which should be applied in any instance:
\begin{itemize}
\item \reproduced The results were validated by a person or team other
than the original authors of the work, with, at least in part,
artifacts provided by the original authors.
\item \replicated As in \reproduced, but without any artifacts
provided by the original authors.
\end{itemize}
Examples:
\begin{itemize}
\item If Asha published a paper with artifacts in 2019, and Tim
published a replication in 2020 using the artifacts, then Asha can
now apply for the \reproduced badge on the 2019 paper.
\item If Cameron published a paper in 2018 with no artifacts, and
Miles published a paper with artifacts in 2020 that independently
obtained the main result, then Cameron can apply for the \replicated
badge on the 2018 paper.
\end{itemize}
If the artifact is accepted as \validated:
\begin{itemize}
\item Authors will be invited to give lightning talks on this work at
the ROSE session at FSE 2021. The ROSE (Recognizing and Rewarding
Open Science in Software Engineering) festival is a world-wide
salute to replication and reproducibility in software engineering.
Our aim is to create a venue where researchers can receive public
credit for facilitating and participating in open science in
software engineering (specifically, in creating replicated and
reproduced results).
\item We will work with the IEEE Xplore and ACM Portal administrator
to add badges to the electronic versions of the paper related to
the artifact.
\end{itemize}
\section{Submission Process}
\label{sec:submission-process}
\subsection{Submission Overview}
In principle, authors are expected to submit through
\href{https://esecfse2021artifacts.hotcrp.com/}{HotCRP} their artifact
documentation. This documentation distinguishes two basic types of
information---captured in one central research abstract (two pages
max)---depending on the intented badge.
\marginlabel{\evaluated:} The emphasis lies on providing documentation
on the research artifact previously prepared and archived. Here, the
authors need to write and submit documentation explaining how to
obtain the artifact package, how to unpack the artifact, how to get
started, and how to use the artifact in more detail. The submission
must only describe the technicalities of the artifact and uses of the
artifact that are not already described in the paper.
\marginlabel{\available:} The authors must give the location of the
artifact on a publicly accessible archival repository, along with a
DOI or a link to the repository. This means that the HotCRP submission
should include the research abstract only providing links to the
repositories where the artifact is permanently stored and available.
Submitting artifacts themselves through HotCRP without making them
publicly accessible (through a repository or an archival service) will
not be sufficient.
\marginlabel{\validated:} The emphasis here lies on providing
information about how their already published research has been
replicated or reproduced as well as links to further material (\eg
the papers and artifacts in question). We encourage submissions for
those badges by the replicating authors nominating the original
authors.
\hypertarget{notavailable}{\attention} If the authors are not aiming
for the \available badge, the artifacts do not necessarily have to be
publicly accessible for the review process. However, \emph{the authors
should clearly explain why the artifact are not publicly available},
for example, because of privacy concerns, law, or NDAs in place. In
this very case, the authors are asked to provide either a private link
/ password-protected link to a repository or they may submit the
artifact directly through HotCRP (in a zip file) and it should become
clear which steps are necessary for authors who would like to reuse
the artifact.
\subsection{Types of Research Artifacts}
There are two options depending on the nature of the artifacts:
\ipackage or \spackage. In general, an \ipackage is related to
software artifacts or, for instance, scripts, while a \spackage may be
related to qualitative studies (\eg interview transcripts or coding
schemas).
In both cases, it is expected that the basic set-up of the artifact
(including configurations and installations) take less than 30
minutes. Otherwise, the artifact is unlikely to be explicitly endorsed
by Program Committee members because they will simply will not have
enough time to deal with it.
\marginlabel{\ipackage}
If the artifact consists of a tool or software system, then the
authors need to prepare an Installation Package so that the tool can
be installed and run in the evaluator’s environment. That is to say,
please make sure to provide enough associated instructions, code, and
data such that any Software Engineering person with a reasonable
knowledge of scripting, build tools, etc., could install, build, and
run the code. If the artifact contains or requires the use of a
special tool or any other non-trivial piece of software, the authors
must provide a VirtualBox VM image or a Docker container image with a
working environment containing the artifact and all the necessary
tools. We expect that the artifacts have been vetted on a clean
machine before submission.
\marginlabel{\spackage}
If the artifact contains documents that can be used with a simple text
editor, a PDF viewer, or some other common tool (\eg a spreadsheet
program in its basic configuration) the authors can just save all
documents in a single package file (zip or tar.gz).
\subsection{Artifact Documentation}
Regardless of the badge, authors must provide documentation explaining
how to obtain the artifact package, how to unpack the artifact, how to
get started, and how to use the artifacts in more detail. The artifact
itself must only describe the technicalities of the artifact and uses
of the artifact that are not already described in the paper;
nevertheless, the artifact and its documentation should be
self-contained. The submission should contain (and / or link to) the
documents listed below. The documents should be in plain text,
MarkDown, or PDF format, indicated by the file extension. The name of
each file should be in capital letters.
\begin{itemize}
\item A \hypertarget{readme}{README} main file describing what the
artifact does and where it can be obtained (with hidden links and
access password if necessary). Also, there should be a clear
description how to repeat, replicate, or reproduce the results
presented in the paper. Artifacts that focus on data should, in
principle, cover aspects relevant to understand the context, data
provenance, ethical and legal statements (as long as relevant), and
storage requirements. Artifacts that focus on software should, in
principle, cover aspects relevant to how to install and use it (and
be accompanied by a small example).
\item A \hypertarget{requirements}{REQUIREMENTS} file for artifacts
that focus on software. This file should, in principle, cover
aspects of hardware environment requirements (\eg performance,
storage or non-commodity peripherals) and software environments (\eg
Docker, VM, and operating system). Ιf relevant, any additional file
with version-specific dependencies information (\eg requirements.txt
for Python-only environments, Cargo.toml for Rust, etc.), should be
included according to the norms of the specific language and
platform. Any deviation from standard environments needs to be
reasonably justified.
\item A \hypertarget{status}{STATUS} file stating what kind of
badge(s) the authors are applying for as well as the reasons why the
authors believe that the artifact deserves that badge(s).
\item A \hypertarget{license}{LICENSE} file describing the
distribution rights. Note that for the \available badge the artifact
needs to be under some form of open source license.
\item An \hypertarget{install}{INSTALL} file with installation
instructions. These instructions should include notes illustrating a
very basic usage example or a method to test the installation. This
could be, for instance, on what output to expect that confirms that
the code is installed and working; and the code is doing something
interesting and useful.
\item A copy of the accepted paper in PDF format.
\end{itemize}
\section{Review Process}
This section's intended audience is the Program Committee and, thus,
addresses the Program Committee members of the Artifact Evaluation
track (and is written accordingly), but it is available to authors as
well to facilitate transparency.
The tasks of the reviewers of research artifacts involve three phases:
\begin{enumerate}
\item Bidding Phase (May 29--June 4, 2021)
\item Initial Review and Rebuttal Phase (June 5--June 19, 2021)
\item In-depth Review Phase (June 19--July 5, 2021)
\end{enumerate}
\subsection{Bidding Phase (May 29--June 4)}
Authors who are planning to submit a research artifact are requested
to register their artifacts by May 28, 2021 using HotCRP. The
submission includes a research abstract with all relevant information
and / or links to the repositories containing the information (such as
the artifact itself). In the exceptional cases described
\hyperlink{notavailable}{above}, the artifact itself may also be
submitted as a zip through HotCRP. For more details, please see the
submission process described above in
\seealso{Section~\ref{sec:submission-process}}
Section~\ref{sec:submission-process}. Immediately after the submission
deadline, we will invite the reviewers to submit their bids in the
HotCRP tool.
\deadline The bidding deadline is June 4, 2021.
Reviewers should consider their conflicts of interest, research
topics, and experiences with specific tools and technologies (if
applicable) when placing their bids.
\subsection{Initial Review and Rebuttal Phase (June 5--June 19)}
Authors will submit their artifacts by June 4, 2021. We will then
assign artifacts to reviewers as soon as possible.
Before the actual In-depth Review Phase (where no interaction with the
authors will take place anymore), reviewers will be asked to check the
integrity of the research artifacts and to look for possible setup
problems or other smaller technical issues that may prevent the
artifact from being properly evaluated (\eg corrupted or missing
files, provided VMs won’t start, immediate crashes on the simplest
example). During this phase, Program Committee members may contact the
authors to request clarifications on the basic installations and
start-up procedures or to resolve simple installation problems.
Reviewers who wish to communicate with the authors of the artifacts
are asked to email the track chairs. In this case we will send the
authors and the reviewers a URL to access a chat allowing them to
\emph{communicate anonymously} during the rebuttal period. The tool we
will use for the communication during the Initial Review and Rebuttal
Phase is Etherpad. The orchestration of the communication is done by
the Program Committee chairs.
To expedite the review process, we are encouraging the reviewers to
try to send all their issues related to installation in one short
message, if possible. Given the short review time available,\emph{ the
authors are expected to respond within a 48-hour period}.
We plan to make any communication between a reviewer and the authors
visible to other reviewers assigned to the same artifact to mitigate
unnecessary overlaps in effort.
\deadline The initial review and rebuttal phase will end on June 19,
2021.
\subsection{In-depth Review Phase (June 20--July 5)}
After the first quick checks during the initial review and rebuttal
phase, possibly leading to the fixing of problems or clarifications
during the initial review and rebuttal phase, the actual in-depth
review will start. We will use a single-blind review process.
Reviewers review the artifact documentation provided by the authors
(\eg referring to the \hyperlink{readme}{README} file in a
repository). \seealso{Section~\ref{sec:submission-process}}
Section~\ref{sec:submission-process} provides further details about
the expected outline of the research artifacts. Except for exceptional
cases, the files comprising the artifact and described in the abstract
should already be publicly accessible through a repository. In
exceptional cases, however, authors might have submitted the files as
a package (e.g. zip) through HotCRP: those cases refer primarily to
the cases where authors do not apply for the \available badge and
where public disclosure of the artifact is not possible, \eg due to
NDAs.
The authors should explain in their submission which badges they are
aiming for (\hyperlink{status}{STATUS} file). The reviewers are
then asked to review the artifact for the respective criteria (see
Section 4) and decide whether the envisioned badge(s) can be awarded,
whether an alternative badge should be awarded (provided the
submission meets the criteria), or whether no badge can be awarded at
all.
Reviewers are expected to assess if and how the things described in
the abstract submission are reflected by the actual artifact in the
repository. However, we would like to stress the importance to avoid a
black and white decision or searching for small issues that prevent
issuing a badge. \emph{The whole point of this track is to promote open
science in our research community and help authors willing to share
their artifacts in doing this correctly (and efficiently)}.
Reviewers are expected to enter the badge decision on HotCRP together
with a short review explaining the badge decision. Please note that we
do not expect an in-depth review report, but only a short explanation
why or why not a certain badge should be awarded. Furthermore, note
that a paper can receive multiple badges.
Artifacts may be awarded one, two, or all three of the \evaluated,
\available, and \validated badges. You can therefore in HotCPR all of
the scores that apply:
\begin{itemize}
\item NO BADGE
\item FUNCTIONAL or REUSABLE
\item AVAILABLE
\item REPRODUCED or REPLICATED
\end{itemize}
\attention Reviewers are asked to submit their reviews as soon as
possible and not to submit all their reviews at once at the end of the
review phase. We allow discussions between reviewers to take place at
any time during the review phase and \emph{all reviews will be made
visible to all reviewers of the same artifact as soon as they are
submitted} to facilitate effective discussions (and feedback/support
by other reviewers) and, again, to mitigate unnecessary overlaps in
effort (\eg to allow reviewers to concentrate on other submissions
first).
Finally, it is allowed to involve an external reviewer in cases the
reviewer would like to obtain additional feedback or expertise. In
that case, it is important to stress the confidentiality of the
process to the external reviewer. However, reviewers are expected to
also familiarize themselves with the research artifact such that they
can assess it fairly. Regardless of the eventual involvement of
external reviewers, please note that \emph{the Program Committee
members assigned to the artifact are personally responsible for the
reviews (with respect to their fairness and accuracy of the
decision)}! Furthermore, we expect the Program Committee members to
personally participate in the online discussion.
\marginlabel{Nominations:} If you want to nominate a research artifact
for the best artifact award, please do so by marking it in the review
form.
\deadline The deadline for submitting reviews is July 5, 2021.
Authors will be notified about the decision on July 9, 2021.
\subsection{Summary of Important Dates}
The timeline for the artifact evaluation track is as follows (All
dates are 23:59:59 AoE (UTC-12h).
\begin{itemize}
\item Friday May 21, 2021: FSE 2021 research paper notification
\item Friday May 28, 2021: Artifact pre-submission registration deadline
\item Saturday May 29, 2021: Artifact Evaluation bidding start
\item Friday June 4, 2021: Artifact Evaluation bidding deadline
\item Friday June 4, 2021: Artifact submission deadline
\item Saturday June 5, 2021: Start of initial review and rebuttal period
\item Saturday June 19, 2021: End of initial review and rebuttal period
\item Sunday June 20, 2021: Start of in-depth review phase
\item Monday July 5, 2021: Artifact Evaluation review submission deadline
\item Friday July 9, 2021: Artifact notification
\end{itemize}
The Artifact Evaluation notification is only about a month before the
conference starts. It is, thus, essential to stick with this schedule!
\section{Evaluation Criteria}
The subsequent checklist contains a non-exhaustive set of criteria for
the evaluation of the artifact submissions for eligibility of the
respective badges. We distinguish minimum criteria (which must be met
to merit receiving the badge) and optional criteria which we
recommend, but do not impose as imperative.
\subsection{\evaluated}
\reusable are an extension of \functional. That is, artifacts which
qualify for \reusable, are per definition \functional but not
necessarily vice-versa. As the scope of the Artifact Evaluation track
is to foster reusability of artifacts, authors are expected to submit
for the \reusable badge.
\subsubsection{Minimum Criteria}
\begin{itemize}
\item[\Checkedbox] Artifacts are well documented and offer, at a
minimum, an inventory of the contents and sufficient description to
enable the artifacts to be exercised.
\item[\Checkedbox] Artifacts are relevant to the associated paper and
contribute to the generation of its main results.
\item[\Checkedbox] Artifacts are self-contained and exercisable and
include scripts and / or software used to generate the results
described in the associated paper, \ie their integrity allows for a
successful execution (if applicable, \ie software-related) and
included data can be accessed and appropriately manipulated.
\item[\Checkedbox] Artifacts have a proper license available for the
artifact, explicitly documented in the
\hyperlink{license}{LICENSE} file.\footnote{The license should
indicate the underlying license model (\eg Creative Commons or
MIT) and potential restrictions. The license text should further
be self-contained (\eg by adding the license text as proposed by,
for example, CC BY to the \hyperlink{license}{LICENCE} file). For
software, we encourage the use of any open source license. For
data, we recommend a Creative Commons license. In any case, the
license should allow reuse for scientific and research purposes.}
\item[\Checkedbox] \ipackages have an explicit documentation of the
requirements / prerequisites necessary for potential installations
or executions of code in the \hyperlink{requirements}{REQUIREMENTS}
file. This also includes requirements towards operating systems and
hardware.
\item[\Checkedbox] \ipackages have an installation script and
step-by-step instructions that allow for the automatic installation
of necessary tools and environments. When environments or operating
systems deviate from the norm (which is essentially always the case
as there is no real norm), the package must include as well virtual
environments (\eg Docker container image or VirtualBox VM image).
The installation must be executable without problems.\footnote{It is
the responsibility of submitting authors to provide an
installation package that allows to run the artifact in the
evaluator’s environment. The instructions themselves should be
kept to the absolutely required minimum and we recommend relying
on virtual environments and automation as much as possible. If the
submission includes a simple package with textual files only (\eg
PDFs or spreadsheets), then these documents can be archived in a
single package (\eg zip or tar.gz). The underlying assumption is
that if artifacts cannot be installed or exercised without
reasonable technical knowledge or without expertise in the
research field, then other authors who would make use of that
artifact may run into problems as well. In this case, we argue,
the badge should not be awarded.}
\end{itemize}
\subsubsection{Optional Criteria}
\begin{itemize}
\item[\Checkedbox] Artifacts have an indication of the time needed to
run them (\eg 1 hour, 4 hours, 2 days) and how to run a shorter
version (\eg 10 min.) to check that it is functional.
\end{itemize}
\attention The identification of potential causes for failed
installations or executions is not part of the reviewers' tasks.
\subsection{\available}
Available artifacts must be made permanently and publicly available,
\ie they are publicly available through a preserved, publicly
accessible repository with a stable URL and a DOI.
\subsubsection{Minimum Criteria}
\begin{itemize}
\item[\Checkedbox] Artifacts are available for public download from a
repository without the need to register.
\item[\Checkedbox] Artifacts are available for public download from a
persistent repository with a stable URL.
\item[\Checkedbox] Each artifact is associated with a Digital Object
Identifier (DOI).
\end{itemize}
\subsubsection{Optional Criteria}
\begin{itemize}
\item[\Checkedbox] Artifacts provide explicit documentation on the
authors and, ideally, instructions and templates on how to cite when
making use of the artifacts. The authors lists are directly
accessible from the main description of the artifact or available
through a dedicated file (\eg AUTHORS).
\end{itemize}
\subsection{\validated}
The criteria for the \validated badges are primarily assessed based on
the submitted research abstracts that outline that (and how) selected
artifacts have reached that stage. That is, reviewers are not expected
to review the actual reproduction entirely and we expect the abstracts
to show that:
\begin{itemize}
\item[\Checkedbox] For \replicated, the main results of the paper have
been obtained in a subsequent study by a person or team other than
the original author, using, in part, the artifacts provided by the
author.
\item[\Checkedbox] For \reproduced, the main results of the paper have
been independently obtained in a subsequent study by a person or
team other than the original authors, without the use of
author-supplied artifacts.
\end{itemize}
The main difference between \replicated and \reproduced lies,
therefore, in whether the external replication (partially) needs to
rely on artifacts by the authors of the research being replicated or
whether the reproduction can be achieved completely independently.
\subsubsection{Minimum Criteria}
\begin{itemize}
\item[\Checkedbox] The paper reporting on the replication /
reproduction has been peer-reviewed.
\item[\Checkedbox] The original paper being reproduced and potentially
awarded the badge is publicly available (via a submitted URL
directory).
\item[\Checkedbox] The authorship of the reproduced / replicated
artifact must not overlap with the reproducing / replicating
artifact.
\item[\Checkedbox] The abstract clearly outlines WHAT is being
reproduced, WHY it is important, and HOW exactly it has been done.
If the replication / reproduction was only partial, then the authors
clearly explain what parts could be achieved or which are missing.
\item[\Checkedbox] The submission lays out substantial evidence for
replication / reproduction.
\item[\Checkedbox] For \replicated only: The abstract clearly shows
that the main results of the paper have been obtained without
author-supplied artifacts.
\end{itemize}
\subsubsection{ Optional Criteria}
\begin{itemize}
\item[\Checkedbox] Authors pay due respect to the other work related
to the reproduction / replication. That is, the abstract is not
necessarily critical towards others in the research community.
\item[\Checkedbox] Mostly only in case the submitting authors are not
the ones of the original work being reproduced / replicated but
authors nominating original work: Authors provide a critical
reflection upon what aspects made it easier or harder to replicate /
reproduce and what are the lessons learned from this work that would
enable more replication / reproduction in the future for other kinds
of tasks or other kinds of research.
\end{itemize}
\subsubsection{Remarks}
To merit the badge \replicated or \reproduced, it is sufficient that
the results are within a margin / tolerance and slightly deviate from
those results of the original study as long as the main claims in the
original paper are not changed. This is especially true for
non-computational studies (\eg qualitative studies). It is not the
responsibility of the reviewers to completely replicate / reproduce
the study by themselves but of the authors to reasonably convey how
this has been achieved. The goal of the Artifact Evaluation track is
to promote work that allows the broader community to use the
artifacts, not in-house specialists only. In case of Reusable
artifacts emerging from, inter alia, more restrictive industrial
research environments, the abstract needs to contain more than
unreproducible claims of the artifact being used, \ie sufficient
details on the actual reproduction / replication to convince the
well-intended reviewers.
\section{Further Supplementary Material to this Document}
While there are various (valuable) contributions related to open
science and, thus, related to this guideline, we recommend the
following supplementary material. Note that the guideline at hands is
intended to be self-contained and the supplementary material is
dedicated to the reader interested in the general notion of open
science.
A broader introduction into the general notion of open science in
Software Engineering, in particular open data and open source which we
consider particularly important to the Artifact Evaluation Track, can
be found in the (open access) book chapter ``Open Science in Software
Engineering'', available here: \href{https://doi.org/fjx4}. This
chapter contains the ABC of open science and pragmatic, short insights
into relevant basics such as proper licensing models.
The recommendations provided in the chapter are also reflected in the
FSE Open Science policy, which we recommend to both reviewers and
authors alike participating in the artifact evaluation track. See also
the \href{https://github.com/acmsigsoft/open-science-policies}{SIGSOFT
Open Science Policies}.
Finally, we recommend the general checklist elaborated by the
Empirical Software Engineering research community as the
\href{https://github.com/acmsigsoft/EmpiricalStandards}{ACM SIGSOFT
Empirical Standards} for researchers, peer reviewers, editors and
publications venues.
\subsubsection{Acknowledgements}
We would like to thank Silvia Abrahão and Daniel Mendez for preparing
the guidelines for the ICSE 2021 Artifact Evaluation track, from which
the present document is derived. Silivia and Daniel would like to
thank Paul Grünbacher and Baishakhi Ray (ICSE 2019 Artifact Evaluation
Co-chairs) for a previous version of a review process outlined in this
guideline. We also want to thank Tim Menzies for his valuable
suggestions and collegial advice as well as Michael Dorner, Alessio
Ferrari, Davide Fucci, and Daniel Graziotin for their valuable
feedback and suggestions on earlier versions of this document.
\end{document}