[HPSG-L] Final cfp: 4REAL 2018 - LREC Workshop on Replicability and Reproducibility
António Branco
antonio.branco at di.fc.ul.pt
Thu Jan 4 09:16:00 UTC 2018
[apologies for duplicates; please help do disseminate]
================= FINAL CALL FOR PAPERS ================
4REAL Workshop 2018
Workshop on Replicability and Reproducibility of Research Results
in Science and Technology of Language
http://4real2018.di.fc.ul.pt
12 May 2018
Miyazaki, Japan
Collocated with LREC 2018
11th Language Resources and Evaluation Conference
http://lrec2018.lrec-conf.org
Important dates:
15 January 2018: Deadline for submissions to 4REAL workshop
9 February 2018: Notification of authors of submissions to 4REAL workshop
1 March 2018: Deadline for camera-ready
9-11 May 2018: LREC main conference (Wednesday to Friday)
12 May 2018: 4REAL workshop (Saturday)
Call for papers:
Reproduction and replication of research results are at the heart of the
validation
of scientific knowledge and of the scientific endeavor. Reproduction of
results
entails arriving at the same overall conclusions that is, to
appropriately validate
a set of results, scientists should strive to reproduce the same answer
to a given
research question by different means, e.g. by reimplementing an algorithm or
evaluating it on a new dataset, etc. Replication has a more limited aim,
typically
involving running the exact same solution or approach under the same
conditions
in order to arrive at the same output result.
Despite their key importance, reproduction and replication have not been
sufficiently
encouraged given the prevailing procedures and priorities for the reviewing,
selection and publication of research results.
The immediate motivation for the increased interest on reproducibility and
replicability is to be found in a number of factors, including the
realization that
for some published results, their replication is not being obtained
(e.g. Prinz et
al., 2011; Belgley and Ellis, 2012); that there may be problems with the
commonly
accepted reviewing procedures, where deliberately falsified submissions,
with
fabricated errors and fake authors, get accepted even in respectable
journals
(e.g. Bohannon, 2013); that the expectation of researchers vis a vis
misconduct,
as revealed in inquiries to scientists on questionable practices, scores
higher than
one might expect or would be ready to accept (e.g. Fanelli, 2009); among
several
others.
This workshop seeks to foster the discussion and the advancement on a
topic that has
been given insufficient attention in the research area of language
processing tools
and resources and that has been an important topic emerging in other
scientific
areas, continuing the objectives of the first edition of the 4REAL
workshop, at LREC
2016. We are thus inviting submissions of articles that present cases,
either with
positive or negative results, of actual replication or reproduction
exercises
of previous published results in our area.
Specific topics within the scope of the call include, but are not
limited to:
– Reproduction and replication of previously published systems, resources,
results, etc.
– Analysis of reproducibility challenges in system-oriented and
user-oriented
evaluation.
– Reproducibility challenges on private or proprietary data.
– Reproducibility challenges on ephemeral data, like streaming data,
tweets, etc.
– Reproducibility challenges on online experiments.
– Reproducibility in evaluation campaigns.
– Evaluation infrastructures and Evaluation as a Service (EaaS).
– Experiments on data management, data curation, and data quality.
– Reproducible experimental workflows: tools and experiences.
– Data citation: citing experimental data, dynamic data sets, samples, etc.
We are interested also in articles discussing the challenges, the risk
factors,
the appropriate procedures, etc. specific to our area or that should be
adopted,
or adapted from other neighboring areas, including methodologies for
monitoring,
maintaining or improving citation of language resources and tools and to
assess
the importance of data citation for research integrity. This includes
also of course
the new risks raised by the replication articles themselves and their
own integrity,
in view of the preservation of the reputation of colleagues and works
whose results
are reported has having been replicated, etc.
“Replicability Focus” in Language Resources and Evaluation Journal:
Best papers will be invited to be submitted to the Language Resources and
Evaluation Journal under its new “Replicability Focus”.
Organization committee:
António Branco (University of Lisbon)
Nicoletta Calzolari (ILC)
Khalid Choukri (ELRA)
Program committee:
Aljoscha Burchardt (DFKI)
António Branco (University of Lisbon)
Gertjan van Noord (University of Groningen)
Joseph Mariani (CNRS/LIMSI)
Khalid Choukri (ELRA)
Maria Gavrilidou (ILSP)
Marko Grobelnik (Jozef Stefan Institute)
Marko Tadic (University of Zagreb)
Nancy Ide (Vassar College)
Nicoletta Calzolari (ILC)
Patrick Paroubek (CNRS/LIMSI))
Piek Vossen (VU University Amsterdam)
Senja Pollak (Jozef Stefan Institute)
Simon Krek (Jozef Stefan Institute)
Stelios Piperidis (ILSP)
Thierry Declerck (DFKI)
Yohei Murakami (Language Grid Japan)
“Identify, Describe and Share your LRs!” initiative:
Describing your Language Resources (LRs) in the LRE Map is now a normal
practice
in the submission procedure of LREC (introduced in 2010 and adopted by other
conferences). To continue the efforts initiated at LREC 2014 about
“Sharing LRs”
(data, tools, web-services, etc.), when submitting a paper, authors will
have
the possibility to upload LRs in a special LREC repository. This effort of
sharing LRs, linked to the LRE Map for their description, may become
a new “regular” feature for conferences in our field, thus contributing
to creating
a common repository where everyone can deposit and share data.
As scientific work requires accurate citations of referenced work so as
to allow
the community to understand the whole context and also replicate the
experiments
conducted by other researchers, LREC 2018 endorses the need to uniquely
Identify LRs
through the use of the International Standard Language Resource Number
(ISLRN,
www.islrn.org), a Persistent Unique Identifier to be assigned to each
Language
Resource. The assignment of ISLRNs to LRs cited in LREC papers will be
offered
at submission time.
References:
Begley and Ellis, 2012, Drug development: Raise standards for preclinical
cancer research, Nature
Bohannon, John, 2013, Who’s Afraid of Peer Review?, Science
Fanelli,2009 How Many Scientists Fabricate and Falsify Research?
Prinz, et al., 2011, Believe it or not: how much can we rely on published
data on potential drug targets?, Nature Reviews Drug Discovery 10, 712.
More information about the HPSG-L
mailing list