posted by user: imDavide || 116 views || tracked by 3 users: [display]

EXTRAAMAS 2025 : 7th International Workshop on EXplainable, Trustworthy, and Responsible AI and Multi-Agent Systems

FacebookTwitterLinkedInGoogle

Link: https://extraamas.ehealth.hevs.ch/
 
When May 19, 2025 - May 20, 2025
Where Detroit
Submission Deadline Mar 1, 2025
Notification Due Mar 15, 2025
Final Version Due Jun 20, 2025
Categories    XAI   explainability   ethics ai   legal ai6th international work
 

Call For Papers

7th International Workshop on
EXplainable and TRAnsparent AI and Multi-Agent Systems
(EXTRAAMAS2025)

in conjunction with AAMAS 202,
Detroit, US, 19-20 May 2025

#Important Dates
Paper submission: 01/03/2025
Notification of acceptance: 15/03/2025
Early registration deadline: TBA
Workshop: 06-07/05/2023
Camera-ready (Springer post-proceedings): 20/06/2025
Submission link: https://easychair.org/conferences/?conf=extraamas2025



Running since 2019, EXTRAAMAS is a well-established workshop and forum on EXplainable and TRAnsparent AI and Multi-Agent Systems. It aims to discuss and disseminate research on explainable artificial intelligence, with a particular focus on intra/inter-agent explainability and cross-disciplinary perspectives. In its 6th edition, EXTRAAMAS identifies four particular focus topics with the ultimate goal of strengthening cutting-edge foundational and applied research. This, of course, comes in addition to the workshop's main theme, focusing, as usual, on XAI fundamentals. The four tracks for this year are:

#Track 1: XAI in symbolic and subsymbolic AI: the “AI dichotomy” separating symbolic AKA classical AI from connectionism AI has persisted for more than seven decades. Nevertheless, the advent of explainable AI has accelerated and intensified the efforts to bridge this gap since providing faithful explanations of black-box machine learning techniques would necessarily mean combining symbolic and subsymbolic AI. This track aims to discuss the recent works on this hot topic of AI.
Track chair: Giovanni Ciatto, University of Bologna, Italy.

#Track 2: XAI in negotiation and conflict resolution: Conflict resolution (e.g., agent-based negotiation, voting, argumentation, etc.) has been a prosperous domain within the MAS community since its foundation. However, as agents and the problems they are tackling become more complex, incorporating explainability becomes vital to assess the usefulness of the supposedly conflict-free solution. This is the main topic of this track, with a special focus on MAS negotiation and explainability.
Track Chair: Reyhan Aydoğan: Ozyegin University, Turkey

#Track 3: Prompts, Interactive Explainability and Dialogues: Appropriate everyday explanations about automated decision-making are context-dependent and interactive. An explanation must fill a 'gap' in the apparent knowledge of the user in a specific context. However, dynamic user modelling is hard. Explanatory dialogue allows designers to try out partial explanations and fine-tune or adjust the explanations based on feedback. This potential for dynamic adjustment can only be redeemed if the system has appropriate interactive capabilities, such as context modelling, user modelling, initiative handling, topic management and grounding. The rapid evolution of LLM and Chatbots has sparked a debate on how to make good use of the interactive capabilities of these new models for explainable AI. The use of LLM also has risks, especially concerning reliability. This triggers relevant methodological questions. How to ensure LLM use reliable data for answering? How to evaluate research based on black-box models? What are good techniques for prompt engineering? In this research track, we welcome new ideas as well as established research outcomes on the wider topic of Interactive or Social Explainable AI.
Track chair: Giovanni Ciatto, University of Bologna

#Track 4: XAI in Law and Ethics: complying with regulation (e.g., GDPR) is among the main objectives for XAI. The right to explanation is key to ensuring transparency of ever more complex AI systems dealing with a multitude of sensitive AI applications. This track discusses works related to explainability in AI ethics, machine ethics, and AI and law.
Track chair: Dr. Rachele Cari, UMEA, & Dr. Simona Tiribelli, UNIMC


#KEYNOTES
This year, EXTRAAMAS will feature two keynote delivered by
Dr. Simona Tiribelli, title: TBA
TBA, title: TBA

All accepted papers are eligible for publication in the Springer Lecture Notes of Artificial Intelligence conference proceedings (after revisions have been applied).


#EXTRAAMAS Tracks
Track1: XAI in symbolic and subsymbolic AI
XAI for Machine learning
Explainable neural networks
Symbolic knowledge injection or extraction
Neuro-symbolic computation
Computational logic for XAI
Multi-agent architectures for XAI
Surrogate models for sub-symbolic predictors
Explainable planning (XAIP)
XAI evaluation

Track2: XAI in negotiation and conflict resolution
Explainable conflict resolution techniques/frameworks
Explainable negotiation protocols and strategies
Explainable recommendation systems
Trustworthy voting mechanisms
Argumentation for explaining the process itself
Argumentation for explaining and supporting the potential outcomes
Explainable user/agent profiling (e.g., learning user's preferences or strategies)
User studies and assessment of the aforementioned approaches
Applications (virtual coaches, robots, IoT)

Track3: Prompts, Interactive Explainability and Dialogue
Interactive capabilities for XAI
Arguments for persuasive explanations
Context modeling
User modeling
Initiative handling
Topic modelling
Grounding and acknowledgment
Prompt engineering
Research methodology for LLM applications
Responsible LLM applications

Track4: (X)AI in Law and Ethics
XAI in AI & Law
Fair AI
XAI & Machine Ethics
Bias reduction
Deception and XAI
Persuasive technologies and XAI
Nudging and XAI
Legal issues of XAI
Liability and XAI
XAI, Transparency, and the Law
Enforceability and XAI
Culture-aware systems and XAI


#Workshop Chairs
Prof. Dr. Davide Calvaresi, HES-SO, Switzerland
research areas: Real-Time Multi-Agent Systems, Explainable AI, BCT, eHealth,
mail: davide.calvaresi@hevs.ch

Dr. Amro Najjar, UNILU, Luxembourg
research areas: Multi-Agent Systems, Explainable AI, AI
mail: amro.najjar@uni.lu

Prof. Dr. Kary Främling, Umeå & Aalto University Sweden/Finland,
research areas: Explainable AI, Artificial Intelligence, Machine Learning, IoT
mail: Kary.Framling@cs.umu.se

Prof. Dr. Andrea Omicini
research areas: Artificial Intelligence, Multi-agent Systems, Soft. Engineering
mail: andrea.omicini@unibo.it, web page, Google Scholar


#Track Chairs
Dr. Giovanni Ciatto, University of Bologna, Italy – giovanni.ciatto@unibo.it
Prof. Rehyan Aydogan, Ozyegin University, Turkey – reyhan.aydogan@ozyegin.edu.tr
Rachele Carli, University of Bologna – rachele.carli2@unibo.it
Joris HULSTIJN: University of Luxembourg – joris.hulstijn@uni.lu

#Advisory Board
Dr. Tim Miller, University of Melbourne
Prof. Dr. Leon van der Torre, UNILU
Prof. Dr. Virginia Dignum, Umea University
Prof. Dr. Michael Ignaz Schumacher

#Primary Contacts
Prof. Dr. Davide Calvaresi - davide.calvaresi@hevs.ch
Dr. Amro Najjar - amro.najjar@list.lu

Related Resources

FLAIRS-ST XAI, Fairness, and Trust 2025   FLAIRS-38 Special Track on Explainable, Fair, and Trustworthy AI
IEEE CITREx 2025   IEEE Symposium on Explainable, Responsible, and Trustworthy Computational Intelligence
EXTRAAMAS 2024   EXplainable and TRAnsparent AI and Multi-Agent Systems
ICAART 2025   17th International Conference on Agents and Artificial Intelligence
AAMAS - 2025   The 24th International Conference on Autonomous Agents and Multiagent Systems
AMSTA 25 2025   19th International Conference on Agents and Multi-Agent Systems: Technology and Applications (AMSTA-25)
KES-IDT 2025   17th International KES Conference on Intelligent Decision Technologies
WMAC 2025   AAAI 2025 Workshop on Advancing LLM-Based Multi-Agent Collaboration
LAMAS&SR 2024   International Workshop on Logical Aspects in Multi-Agent Systems and Strategic Reasoning
MALTA 2025   AAAI-2025 Workshop: Multi-Agent Reinforcement Learning for Transportation Autonomy