[Corpora-List] CFP: ACM RecSys 2013 Workshop on Reproducibility and Replication in Recommender Systems Evaluation (RepSys 2013)
Alejandro Bellogin Kouki
alejandro.bellogin at uam.es
Mon Jun 10 17:09:00 UTC 2013
[Apologies if you receive this more than once]
===============================================================================
ACM RecSys Workshop on
Reproducibility and Replication in Recommender Systems Evaluation - RepSys2013
7th ACM Recommender Systems Conference (RecSys 2013)
Hong Kong, China, 12 or 16 October 2013
http://repsys.project.cwi.nl
===============================================================================
* Submission deadline: 22 July 2013 *
== Scope ==
Experiment replication and reproduction are key requirements for empirical
research methodology, and an important open issue in the field of Recommender
Systems. When an experiment is repeated by a different researcher and exactly
the same result is obtained, we can say the experiment has been replicated. When
the results are not exactly the same but the conclusions are compatible with the
prior ones, we have a reproduction of the experiment. Reproducibility and
replication involve recommendation algorithm implementations, experimental
protocols, and evaluation metrics. While the problem of reproducibility and
replication has been recognized in the Recommender Systems community, the need
for a clear solution remains largely unmet, which motivates the present
workshop.
== Topics ==
We invite the submission of papers reporting original research, studies,
advances, experiences, or work in progress in the scope of reproducibility and
replication in Recommender Systems evaluation. Papers explicitly dealing with
replication of previously published experimental conditions/algorithms/metrics
and the resulting analysis are encouraged. In particular, we seek discussions on
the difficulties the authors may find in this process, along with their
limitations or successes on reproducing the original results.
The topics the workshop seeks to address include –though need not be limited to–
the following:
* Limitations and challenges of experimental reproducibility and replication
* Reproducible experimental design
* Replicability of algorithms
* Standardization of metrics: definition and computation protocols
* Evaluation software: frameworks, utilities, services
* Reproducibility in user-centric studies
* Datasets and benchmarks
* Recommender software reuse
* Replication of already published work
* Reproducibility within and across domains and organizations
* Reproducibility and replication guidelines
== Submission ==
Two submission types are accepted: long papers of up to 8 pages, and short
papers up to 4 pages. The papers will be evaluated for their originality,
contribution significance, soundness, clarity, and overall quality. The interest
of contributions will be assessed in terms of technical and scientific findings,
contribution to the knowledge and understanding of the problem, methodological
advancements, or applicative value. Specific contributions focusing on
repeatability and reproducibility in terms of algorithm implementations,
evaluation frameworks and/or practice will also be welcome and valued.
All submissions shall adhere to the standard ACM SIG proceedings format:
http://www.acm.org/sigs/publications/proceedings-templates.
Submissions shall be sent as a pdf file through the online submission system now
open at: https://www.easychair.org/conferences/?conf=repsys2013.
== Important dates ==
* Paper submission deadline: 22 July
* Notification: 16 August
* Camera-ready version due: 30 August
== Organizers ==
* Alejandro Bellogín, Centrum Wiskunde & Informatica, The Netherlands
* Pablo Castells, Universidad Autónoma de Madrid, Spain
* Alan Said, Centrum Wiskunde & Informatica, The Netherlands
* Domonkos Tikk, Gravity R&D, Hungary
== Programme Committee ==
* Xavier Amatriain, Netflix, USA
* Linas Baltrunas, Telefonica Research, Spain
* Marcel Blattner, University of Applied Sciences, Switzerland
* Iván Cantador, Universidad Autónoma de Madrid, Spain
* Ed Chi, Google, USA
* Arjen de Vries, Centrum Wiskunde & Informatica, Netherlands
* Juan Manuel Fernández, Universidad de Granada, Spain
* Zeno Gantner, Nokia, Germany
* Pankaj Gupta, Twitter, USA
* Andreas Hotho, University of Würburg, Germany
* Juan Huete, Universidad de Granada, Spain
* Kris Jack, Mendeley, England
* Dietmar Jannach, University of Dortmund, Germany
* Jaap Kamps, University of Amsterdam, Netherlands
* Alexandros Karatzoglou, TID, Spain
* Bart Knijnenburg, University of California, Irvine, USA
* Ido Guy, Google, Israel
* Jérôme Picault, Bell Labs, Alcatel-Lucent, France
* Till Plumbaum, TU Berlin, Germany
* Daniele Quercia, Yahoo!, Spain
* Filip Radlinski, Microsoft, Canada
* Yue Shi, TU-Delft, The Netherlands
* Fabrizio Silvestri, Consiglio Nazionale delle Ricerche, Italy
* Harald Steck, Netflix, USA
* David Vallet, NICTA, Australia
* Jun Wang, University College London, UK
* Xiaoxue Zhao, University College London, UK
_______________________________________________
UNSUBSCRIBE from this page: http://mailman.uib.no/options/corpora
Corpora mailing list
Corpora at uib.no
http://mailman.uib.no/listinfo/corpora
More information about the Corpora
mailing list