posted by user: grupocole || 395 views || tracked by 1 users: [display]

L2M 2025 : First Large Language Model Memorization (L2M2) workshop @ACL 2025

FacebookTwitterLinkedInGoogle

 
When Jul 31, 2025 - Aug 1, 2025
Where Vienna, Austria
Submission Deadline Mar 25, 2025
Notification Due Apr 27, 2024
Final Version Due May 16, 2025
Categories    NLP   computational linguistics   artificial intelligence
 

Call For Papers

Call for Papers for the L2M2 workshop
Large language models (LLMs) are known to memorize their training data. In recent years, this phenomenon has inspired multiple distinct research directions. Some researchers currently focus on analyzing and understanding LLM memorization, attempting to localize memorized knowledge within a model or understand which examples are most likely to be memorized. Other researchers aim to edit or remove information that an LLM has memorized. Still, others study the downstream implications of LLM memorization, including legal concerns associated with memorizing copyrighted articles, privacy risks associated with LLMs leaking private information, and concerns that LLMs cheat on benchmarks by memorizing test data.

The First Workshop on Large Language Model Memorization (L2M2), co-located with ACL 2025 in Vienna, seeks to provide a central venue for researchers studying LLM memorization from these different angles. We invite paper submissions on all topics related to LLM memorization, including but not limited to:

Behavioral analyses that seek to understand what training data is memorized by models.

Methods for measuring the extent to which models memorize training data.

Interpretability work analyzing how models memorize training data.

The relationship between training data memorization and membership inference.

Analyses of how memorization of training data is related to generalization and model capabilities.

Methods or benchmarks for preventing models from outputting memorized data, such as machine unlearning, model editing, or output filtering techniques.

Model editing techniques for modifying knowledge that models have memorized.

Legal implications and risks of LLM memorization.

Privacy and security risks of LLM memorization.

Implications of memorization for benchmarking, such as data contamination concerns.

Any models that process or generate text are in scope for this workshop, including text-only language models, vision language models, machine translation models, etc. Memorization in the context of L2M2 could refer to verbatim memorization of training data, approximate memorization of copyrighted content, memorization of factual knowledge, memorization of correct answers to questions, or other definitions.

If you have any questions about whether your paper is suitable for our workshop, feel free to contact us at l2m2-workshop@googlegroups.com.

Submission Guidelines
We seek submissions of at least 4 and at most 8 pages, not including references. All submissions will be reviewed in a single track, regardless of length. Please format your papers using the standard ACL format. We will accept direct submissions via OpenReview, as well as submissions made through ACL Rolling Review.

We accept two types of submissions:

Archival: Original work that will be included in the workshop proceedings. Submissions must be anonymized and will go through a double-blind review process.

Non-archival: Such submissions will not be included in the proceedings. The workshop will accept two categories of non-archival work:

Work published elsewhere: The authors should indicate the original venue, and will be accepted if the organizers think the work will benefit from exposure to the audience of this workshop. Submissions do not need to be anonymized.

Work-in-progress: These may report on work in progress that have not yet been published at a different venue. The acceptance will be based on a double-blind review and thus submissions of this type need to be anonymized.

Dual submissions
We allow submissions that are also under review in other venues, but please note that many conferences do not allow it, so make sure that you do not violate their policy as well.

Anonymity period
We do not have an anonymity deadline. Preprints are allowed, both before the submission deadline as well as after.

Important Dates

Latest submission date for ARR submissions: February 15, 2025

Direct submission deadline: March 25, 2025

Pre-reviewed (ARR) commitment deadline: April 17, 2025

Notification of acceptance: April 27, 2025

Camera-ready paper deadline: May 16, 2025

Proceedings due (hard deadline): June 30, 2025

Pre-recorded video due (hard deadline): July 7, 2025

Workshop dates: July 31st - August 1st 2025

Note: all deadlines are 11:59PM UTC-12:00.

The organisers:
Robin Jia

Eric Wallace

Yangsibo Huang

Tiago Pimentel

Pratyush Maini

Verna Dankers

Johnny Wei

Pietro Lesci

Related Resources

IEEE-Ei/Scopus-ITCC 2025   2025 5th International Conference on Information Technology and Cloud Computing (ITCC 2025)-EI Compendex
LLM4SE 2025   1st Workshop on Large Language Models for Generative Software Engineering
ACL 2025   The 63rd Annual Meeting of the Association for Computational Linguistics
GenAI and LVMs for Biometrics 2025   IEEE Transactions on Biometrics, Behavior, and Identity Science (T-BIOM) Special Issue on Generative AI and Large Vision-Language Models for Biometrics
CoNLL 2025   29th Conference on Computational Natural Language Learning
SPIE-Ei/Scopus-DMNLP 2025   2025 2nd International Conference on Data Mining and Natural Language Processing (DMNLP 2025)-EI Compendex&Scopus
LDK 2025   Fifth Conference on Language, Data and Knowledge
ISWC 2025   24th International Semantic Web Conference
IEILM 2025   The 2nd Workshop on Integrating Edge Intelligence and Large Model in Next Generation Networks
MAD 2025   Call For Papers: 4th ACM International Workshop on Multimedia AI against Disinformation