32.3855, FYI: Call for Participation: SemEval-2022 Task 09: R2VQ
The LINGUIST List
linguist at listserv.linguistlist.org
Wed Dec 8 20:19:35 UTC 2021
LINGUIST List: Vol-32-3855. Wed Dec 08 2021. ISSN: 1069 - 4875.
Subject: 32.3855, FYI: Call for Participation: SemEval-2022 Task 09: R2VQ
Moderator: Malgorzata E. Cavar (linguist at linguistlist.org)
Student Moderator: Jeremy Coburn, Lauren Perkins
Managing Editor: Becca Morris
Team: Helen Aristar-Dry, Everett Green, Sarah Robinson, Nils Hjortnaes, Joshua Sims, Billy Dickson
Jobs: jobs at linguistlist.org | Conferences: callconf at linguistlist.org | Pubs: pubs at linguistlist.org
Homepage: http://linguistlist.org
Please support the LL editors and operation with a donation at:
https://funddrive.linguistlist.org/donate/
Editor for this issue: Everett Green <everett at linguistlist.org>
================================================================
Date: Wed, 08 Dec 2021 15:19:01
From: James Pustejovsky [pustejovsky at gmail.com]
Subject: Call for Participation: SemEval-2022 Task 09: R2VQ
FIRST CALL FOR PARTICIPATION
SemEval-2022 Task 09: R2VQ - Competence-based Multimodal Question Answering
We invite you to participate in the SemEval-2022 Task 9: Competence-based
Multimodal Question Answering (R2VQ).
The task is being held as part of SemEval-2022, and all participating team
will be able to publish their system description paper in the proceedings
published by ACL.
Codalab (Data download): https://competitions.codalab.org/competitions/34056
Motivation
================================================
When we apply our existing knowledge to new situations, we demonstrate a kind
of understanding of how the knowledge (through tasks) is applied. When viewed
over a conceptual domain, this constitutes a competence. Competence-based
evaluations can be seen as a new approach for designing NLP challenges, in
order to better characterize the underlying operational knowledge that a
system has for a conceptual domain, rather than focusing on individual tasks.
In this shared task, we present a challenge that is reflective of linguistic
and cognitive competencies that humans have when speaking and reasoning.
Task Overview
================================================
Given the intuition that textual and visual information mutually inform each
other for semantic reasoning, we formulate the challenge as a
competence-based question answering (QA) task, designed to involve rich
semantic annotation and aligned text-video objects. The task is structured as
question answering pairs, querying how well a system understands the semantics
of
recipes.
We adopt the concept of ''question families'' as outlined in the CLEVR dataset
(Johnson et al., 2017). While some question families naturally transfer over
from the VQA domain (e.g., integer comparison, counting), other concepts such
as ellipsis and object lifespan must be employed to cover the full extent of
competency within procedural texts.
Data Content
================================================
We have built the R2VQ (Recipe Reading and Video Question Answering) dataset,
a dataset consisting of a collection of recipes sourced from
https://recipes.fandom.com/wiki/Recipes_Wiki and foodista.com, and labeled
according to three distinct annotation layers: (i) Cooking Role Labeling
(CRL), (ii) Semantic Role Labeling (SRL), and (iii) aligned image frames taken
from creative commons cooking videos downloaded from YouTube. It consists of
1,000 recipes, with 800 to be used as training, and 100 recipes each for
validation and testing. Participating systems will be exposed to the
aforementioned multimodal training set, and will be asked to provide answers
to unseen queries exploiting (i) visual and textual information jointly, or
(ii) textual information only.
Task Website and Codalab Submission site:
https://competitions.codalab.org/competitions/34056
Mailing List: semeval-2022-task9 at googlegroups.com
Important Dates
================================================
Training data available: October 15, 2021
Validation data available: December 3, 2021
Evaluation data ready: December 3, 2021
Evaluation start: January 10, 2021
Evaluation end: January 31, 2022
System Description Paper submissions due: February 23, 2022
Notification to authors: March 31, 2022
Organization
================================================
James Pustejovsky, Brandeis University, jamesp at brandeis.edu
Jingxuan Tu, Brandeis University, jxtu at brandeis.edu
Marco Maru, Sapienza University of Rome, maru at di.uniroma1.it
Simone Conia, Sapienza University of Rome, conia at di.uniroma1.it
Roberto Navigli, Sapienza University of Rome, navigli at diag.uniroma1.it
Kyeongmin Rim, Brandeis University, krim at brandeis.edu
Kelley Lynch, Brandeis University, kmlynch at brandeis.edu
Richard Brutti, Brandeis University, richardbrutti at brandeis.edu
Eben Holderness, Brandeis University, egh at brandeis.edu
Linguistic Field(s): Computational Linguistics
------------------------------------------------------------------------------
*************************** LINGUIST List Support ***************************
The 2020 Fund Drive is under way! Please visit https://funddrive.linguistlist.org
to find out how to donate and check how your university, country or discipline
ranks in the fund drive challenges. Or go directly to the donation site:
https://crowdfunding.iu.edu/the-linguist-list
Let's make this a short fund drive!
Please feel free to share the link to our campaign:
https://funddrive.linguistlist.org/donate/
----------------------------------------------------------
LINGUIST List: Vol-32-3855
----------------------------------------------------------
More information about the LINGUIST
mailing list