32.3855, FYI: Call for Participation: SemEval-2022 Task 09: R2VQ

The LINGUIST List linguist at listserv.linguistlist.org
Wed Dec 8 20:19:35 UTC 2021


LINGUIST List: Vol-32-3855. Wed Dec 08 2021. ISSN: 1069 - 4875.

Subject: 32.3855, FYI: Call for Participation: SemEval-2022 Task 09: R2VQ

Moderator: Malgorzata E. Cavar (linguist at linguistlist.org)
Student Moderator: Jeremy Coburn, Lauren Perkins
Managing Editor: Becca Morris
Team: Helen Aristar-Dry, Everett Green, Sarah Robinson, Nils Hjortnaes, Joshua Sims, Billy Dickson
Jobs: jobs at linguistlist.org | Conferences: callconf at linguistlist.org | Pubs: pubs at linguistlist.org

Homepage: http://linguistlist.org

Please support the LL editors and operation with a donation at:
           https://funddrive.linguistlist.org/donate/

Editor for this issue: Everett Green <everett at linguistlist.org>
================================================================


Date: Wed, 08 Dec 2021 15:19:01
From: James Pustejovsky [pustejovsky at gmail.com]
Subject: Call for Participation: SemEval-2022 Task 09: R2VQ

 
FIRST CALL FOR PARTICIPATION

SemEval-2022 Task 09: R2VQ - Competence-based Multimodal Question Answering

We invite you to participate in the SemEval-2022 Task 9: Competence-based
Multimodal Question Answering (R2VQ).

The task is being held as part of SemEval-2022, and all participating team
will be able to publish their system description paper in the proceedings
published by ACL.

Codalab (Data download): https://competitions.codalab.org/competitions/34056

Motivation 
================================================
When we apply our existing knowledge to new situations, we demonstrate a kind
of understanding of how the knowledge (through tasks) is applied. When viewed
over a conceptual domain, this constitutes a competence. Competence-based
evaluations  can be  seen as a new approach for designing NLP challenges, in
order to better characterize the underlying operational knowledge that a
system has for a  conceptual domain, rather than focusing on individual tasks.
In this shared task, we present a challenge that is reflective of linguistic
and cognitive competencies that humans have when speaking and reasoning.

Task Overview 
================================================
Given the intuition that textual and visual information mutually inform each
other for semantic reasoning, we formulate the  challenge as a
competence-based question answering (QA) task, designed to involve rich
semantic annotation and aligned text-video objects. The task is structured as
question answering pairs, querying how well a system understands the semantics
of
recipes.

We adopt the concept of ''question families'' as outlined in the CLEVR dataset
(Johnson et al., 2017). While some question families naturally transfer over
from the VQA domain (e.g., integer comparison, counting), other concepts such
as ellipsis and object lifespan must be employed to cover the full extent of
competency within procedural texts. 

Data Content
================================================ 
We have built the R2VQ (Recipe Reading and Video Question Answering) dataset,
a dataset consisting of a collection of recipes sourced from
https://recipes.fandom.com/wiki/Recipes_Wiki and foodista.com, and labeled
according to three distinct annotation layers: (i) Cooking Role Labeling
(CRL), (ii) Semantic Role Labeling (SRL), and (iii) aligned image frames taken
from creative commons cooking videos downloaded from YouTube. It consists of
1,000 recipes, with 800 to be used as training, and 100 recipes each for
validation and testing. Participating systems will be exposed to the
aforementioned multimodal training set, and will be asked to provide answers
to unseen queries exploiting (i) visual and textual information jointly, or
(ii) textual information only. 

Task Website and Codalab Submission site:
https://competitions.codalab.org/competitions/34056
Mailing List: semeval-2022-task9 at googlegroups.com

Important Dates
================================================ 
Training data available: October 15, 2021

Validation data available: December 3, 2021

Evaluation data ready: December 3, 2021

Evaluation start: January 10, 2021

Evaluation end: January 31, 2022

System Description Paper submissions due: February 23, 2022

Notification to authors: March 31, 2022
 
 
Organization 
================================================ 
James Pustejovsky, Brandeis University, jamesp at brandeis.edu 
Jingxuan Tu, Brandeis University, jxtu at brandeis.edu 
Marco Maru, Sapienza University of Rome, maru at di.uniroma1.it 
Simone Conia, Sapienza University of Rome, conia at di.uniroma1.it 
Roberto Navigli, Sapienza University of Rome, navigli at diag.uniroma1.it 
Kyeongmin Rim, Brandeis University, krim at brandeis.edu
Kelley Lynch, Brandeis University, kmlynch at brandeis.edu 
Richard Brutti, Brandeis University, richardbrutti at brandeis.edu 
Eben Holderness, Brandeis University, egh at brandeis.edu
 



Linguistic Field(s): Computational Linguistics





 



------------------------------------------------------------------------------

***************************    LINGUIST List Support    ***************************
 The 2020 Fund Drive is under way! Please visit https://funddrive.linguistlist.org
  to find out how to donate and check how your university, country or discipline
     ranks in the fund drive challenges. Or go directly to the donation site:
                   https://crowdfunding.iu.edu/the-linguist-list

                        Let's make this a short fund drive!
                Please feel free to share the link to our campaign:
                    https://funddrive.linguistlist.org/donate/
 


----------------------------------------------------------
LINGUIST List: Vol-32-3855	
----------------------------------------------------------






More information about the LINGUIST mailing list