36.3681, Confs: SemEval-2026 Workshop Task 13: Detecting Machine-Generated Code (Online)
The LINGUIST List
linguist at listserv.linguistlist.org
Fri Nov 28 12:05:02 UTC 2025
LINGUIST List: Vol-36-3681. Fri Nov 28 2025. ISSN: 1069 - 4875.
Subject: 36.3681, Confs: SemEval-2026 Workshop Task 13: Detecting Machine-Generated Code (Online)
Moderator: Steven Moran (linguist at linguistlist.org)
Managing Editor: Valeriia Vyshnevetska
Team: Helen Aristar-Dry, Mara Baccaro, Daniel Swanson
Jobs: jobs at linguistlist.org | Conferences: callconf at linguistlist.org | Pubs: pubs at linguistlist.org
Homepage: http://linguistlist.org
Editor for this issue: Valeriia Vyshnevetska <valeriia at linguistlist.org>
================================================================
Date: 27-Nov-2025
From: Preslav Nakov [preslav.nakov at gmail.com]
Subject: SemEval-2026 Workshop Task 13: Detecting Machine-Generated Code
SemEval-2026 Workshop Task 13: Detecting Machine-Generated Code
Short Title: SemEval-2026 Task 13
Location: Online
Meeting URL: https://github.com/mbzuai-nlp/SemEval-2026-Task13
Linguistic Field(s): Computational Linguistics
Call for participation:
SemEval-2026 Task 13: Detecting Machine-Generated Code
We announce an exciting and challenging new SemEval task!
With the rapid growth of large language models for code generation, it
is becoming increasingly difficult to distinguish between code written
by humans and code produced by AI systems. This raises serious
concerns for academic integrity, hiring evaluations, and software
security.
To address this, we are introducing SemEval-2026 Task 13: Detecting
Machine-Generated Code with Multiple Programming Languages,
Generators, and Application Scenarios. It is focused on building
systems that can tell whether code was authored by humans or by LLMs.
The task includes three subtasks:
Subtask A: Binary Classification (Human vs. Machine)
https://www.kaggle.com/t/99673e23fe8546cf9a07a40f36f2cc7e
A real-world challenge with strong out-of-distribution shifts in the
test set. Participants will need to design robust models capable of
adapting to unseen programming languages and contexts: for instance,
determining whether a system trained on competitive programming data
can generalize to identifying LLM-authore code in GitHub repositories.
Subtask B: Multiclass LLM Authorship Identification
https://www.kaggle.com/t/65af9e22be6d43d884cfd6e41cad3ee4
Participants must identify which specific LLM produced a given code
snippet. This subtask aims to explore how stylistic and structural
differences across models can support source attribution.
Subtask C: Four-Way Hybrid Authorship Classification
https://www.kaggle.com/t/005ab8234f27424aa096b7c00a073722
A 4-class classification task task. We ask participants to build a
system which can not only identify fully haman or LLM-generated codes,
but also to find codes with hybrid authorship (human written and then
LLM adjusted) and more seriously, codes that were generated by LLMs in
attempt to fool detectors (either by prompting or by RLHF
fine-tuning).
If you're interested in tackling one of the most relevant and
high-impact challenges in code forensics and AI transparency, explore
our GitHub repository:
https://github.com/mbzuai-nlp/SemEval-2026-Task13
------------------------------------------------------------------------------
********************** LINGUIST List Support ***********************
Please consider donating to the Linguist List, a U.S. 501(c)(3) not for profit organization:
https://www.paypal.com/donate/?hosted_button_id=87C2AXTVC4PP8
LINGUIST List is supported by the following publishers:
Bloomsbury Publishing http://www.bloomsbury.com/uk/
Cambridge University Press http://www.cambridge.org/linguistics
Cascadilla Press http://www.cascadilla.com/
De Gruyter Brill https://www.degruyterbrill.com/?changeLang=en
Edinburgh University Press http://www.edinburghuniversitypress.com
John Benjamins http://www.benjamins.com/
Language Science Press http://langsci-press.org
Lincom GmbH https://lincom-shop.eu/
MIT Press http://mitpress.mit.edu/
Multilingual Matters http://www.multilingual-matters.com/
Narr Francke Attempto Verlag GmbH + Co. KG http://www.narr.de/
Netherlands Graduate School of Linguistics / Landelijke (LOT) http://www.lotpublications.nl/
Peter Lang AG http://www.peterlang.com
----------------------------------------------------------
LINGUIST List: Vol-36-3681
----------------------------------------------------------
More information about the LINGUIST
mailing list