posted by user: grupocole || 801 views || tracked by 2 users: [display]

STAND4NP 2024 : STAND Workshop on Standardizing Tasks, meAsures and NLP Datasets

FacebookTwitterLinkedInGoogle

Link: https://stand4nlp.github.io/
 
When Jan 29, 2024 - Jan 29, 2024
Where Paris, France
Submission Deadline Jan 24, 2024
Categories    NLP   computational linguistics   artificial intelligene
 

Call For Papers


STAND Workshop on Standardizing Tasks, meAsures and NLP Datasets
https://stand4nlp.github.io/

Full-day workshop in Paris, France, January 29th 2024 (+ partial hybrid)

Abstract submission deadline: January 24th 2024, but earlier submissions are welcome

Scientific context:

The current lack of standardized practices and definitions in NLP systems hinders the progress of the field. Indeed, there is not always consensus on which evaluation methods are meaningful and fruitful, or which of their implementations are to be used with which parameters (eg. SacreBLEU, Post 2018).
In some cases, there is no general agreement on the very definition of a task.
This situation calls for work on standardizing NLP practices.

The International Organization for Standardization (ISO) has just created a dedicated working group on NLP (as a joint effort of the AI and Language committees), and 2 standards are already under way. Topics under consideration by the ISO standardization committees include NLP terminology, evaluation metrics, interoperability, annotation guidelines, good practices in NLP development/evaluation/corpora, documentation.

These topics are already heavily discussed in academia, and a number of informal guidelines have already been proposed. We believe that the creation of NLP standards can significantly benefit from the input of both NLP academics and industry NLP practitioners.
Reciprocally, NLP researchers would benefit from getting involved in the standardization effort, thus ensuring that academia's views are listened to, in particular in the context of the AI Act (the European regulation on AI that has been finalized in December), whose enforcement will strongly rely on those standards.

The STAND workshop is a research initiative whose goal is:

to foster discussion on existing standards, their creation and use
to assess the current needs of the community for standardization
to share experience on the impact on the research activities when lacking good practices
to collect existing good practices (and propose new ones)


We invite contributions from NLP practitioners from both the industry and academia, as well as standardization experts.

We invite two types of submission:
* short abstract: 1 page
* long abstract: 3 pages

Accepted submissions will be presented as posters. Authors accepted in the long-abstract track will be invited to submit a full paper (5-10 pages) after the workshop.
Topics for submissions include, but are not limited to:

Comparability and reproducibility of evaluation setup
Annotation guidelines
Evaluation metrics
Good practices for building, annotating and maintaining corpora
Good practices for system evaluation
Interoperability
Ethical guidelines
Guidelines for documenting corpora and models


Submission instructions:

Submissions are expected in PDF form by email at stand4nlp@inria.fr
All submissions should be formatted using the ACL 2023 style files https://2023.aclweb.org/calls/style_and_formatting/.


============

PROGRAM AT A GLANCE:

[09:00-10:00] Welcome, introduction to standardization, ongoing activities in NLP standardization, and the AI Act context
[10:15-11:50] Academic keynote (Joakim Nivre) and invited talks (Matt Post, other speaker TBC)
[11:50-13:30] Poster session (with boosters) & lunch
[13:30-14:40] Industry keynote (speaker TBC) and invited talk (Dirk Hovy)
[15:00-16:30] Moderator-led breakout discussions. Potential topics that will be discussed include:
- [sharing / drafting] Standardizing good practices for evaluation
- [sharing / drafting] Standardizing good practices for corpus management (collection, annotation, versioning)
- [sharing / drafting] Standardizing evaluation metrics (definitions, implementation, sharing scripts)
- [sharing / drafting] Standardizing annotation schemes (formats and guidelines)
- [debate] Explainability and ethics in NLP: what needs for standards?
- [debate] Comparing standardization needs with limitations of the state-of-the-art: how to bridge the gap?
- [debate] Towards standardizing translations of technical terminology in NLP: how to organize i18n?
[16:30-17:30] Reports from breakouts, definition of community-level actions & wrap-up. Example outcomes that are envisioned include:
- Collection and drafting of existing good practices
- Preparation of a joint submission for a position paper
- Creation of common repositories for evaluation scripts, corpus documentation


Participants to the workshop will be offered the opportunity to attend a standardization committee's meeting, which has been scheduled for the day after the workshop (January 30th). The outputs of that meeting will be used in direct support of the AI Act.

Remote access will be offered for part of the workshop only. In-person participation is recommended if possible.
Posters will be in-person only.


IMPORTANT DATES:
Abstract submission: Anytime by January 24
Notification of acceptance: Within a few days of submission
Workshop: January 29
Standardization committee meeting: January 30


ORGANISING COMMITTEE:
Lauriane Aufrant, Timothée Bernard, Maximin Coavoux, Yoann Dupont, Arnaud Ferré, Taras Holoyad, Rania Wazir

MORE INFORMATION
For the latest information see the workshop page at
https://stand4nlp.github.io/; for any questions contact stand4nlp@inria.fr.

Related Resources

IEEE-Ei/Scopus-ITCC 2025   2025 5th International Conference on Information Technology and Cloud Computing (ITCC 2025)-EI Compendex
ISIPTA 2025   14th International Symposium on Imprecise Probabilities
COIT 2025   5th International Conference on Computing and Information Technology
ACL 2025   The 63rd Annual Meeting of the Association for Computational Linguistics
BIOEN 2025   8th International Conference on Biomedical Engineering and Science
LSIJ 2024   Life Sciences: an International Journal
AISCA 2025   International Conference on Artificial Intelligence, Soft Computing And Applications
PREDICTION SOLUTIONS 2025   International Conference on Prediction Solutions for Technical and Societal Systems
NMCO 2025   11th International Conference on Networks, Mobile Communication