Workshops
GENEA: Generation and Evaluation of Non-verbal Behaviour for Embodied Agents
Click here to go to the workshop site
The GENEA (Generation and Evaluation of Non-verbal Behaviour for Embodied Agents) Workshop 2026 aims to bring together researchers from diverse disciplines working on different aspects of non-verbal behaviour generation, facilitating discussions on advancing both generation techniques and evaluation methodologies. We invite contributions from fields such as human-computer interaction, machine learning, multimedia, robotics, computer graphics, and social sciences.
Organisers:
- Taras Kucherenko (Electronic Arts, Sweden)
- Alice Delbosc (DAVI-Les Humaniseurs,France)
- Gustav Eje Henter (KTH Royal Institute of Technology, Sweden)
- Oya Celiktutan (King’s College London, UK)
- Eneko Atxa Landa (University of the Basque Country, Spain)
- Jieyeon Woo (Korea Institute of Machinery and Materials, South Korea)
- Haoyang Du (Technological University Dublin, Ireland)
—
Beyond a Concrete: Workshop on Grounding Abstract Concepts in Multimodal Interaction
Click here to go to the workshop site
Abstract concepts pose a fundamental challenge for multimodal interactive systems, as they cannot be grounded in single perceptual features or fixed motor patterns; instead, they require abstraction across perception, action, language, and affect. Examples include superordinate categories (e.g., animal, tool), generalised actions (e.g., use, make), and evaluative concepts (e.g., good, appropriate) that depend on the interaction context. This workshop focuses on computational and interactional mechanisms for grounding abstract concepts beyond concrete sensory inputs, bringing together work on multimodal representation learning, embodied and developmental models, robotics, and hybrid cognitive architectures. The workshop aims to consolidate approaches, datasets, and evaluation strategies for abstract concept grounding, identify shared modelling assumptions, and clarify open challenges. Expected outcomes include a comparative discussion of architectures and benchmarks and a post-workshop summary outlining research gaps and future directions, supporting multimodal interactive systems with improved generalisation and interpretability.
Organisers:
- Rahul Singh Maharjan (University of Manchester, UK)
- Haodong Xie (University of Manchester, UK)
- Niyati Rawal (BITS-Goa, India)
- Luca Raggioli (University of Naples Federico II, Italy)
- Angelo Cangelosi (University of Manchester, UK)
—
Inclusive AI: Rethinking AI-based Multimodal Interaction for Diverse and Underrepresented Users
Click here to go to the workshop site
This workshop marks the inaugural edition of “Inclusive AI”, an initiative focused on addressing social interaction challenges in AI systems, with a particular emphasis on inclusivity for diverse and underrepresented user populations. It builds on a growing recognition across the HCI, HRI, and AI communities of the need for cross-disciplinary venues to advance socially interactive AI that is equitable, adaptive, and user-aware.
Organisers:
- Giulia Barbareschi (University of Duisburg Essen)
- Nataliya Kosmyna (MIT Media Lab and Google)
- Maristella Matera (Politecnico di Milano)
- Anoop Sinha (Google)
- Shruti Sheth (Google)
- Micol Spitale (Politecnico di Milano)
- Alessandro Vinciarelli (University of Glasgow)
—
ACCESS-MI: Context-Aware Assistive Agents for Accessible Computing
Click here to go to the workshop site
Many blind and visually impaired individuals still struggle to use computers due to the increasing complexity of modern-day user interfaces. Screen-reader workflows often break and require slow, unreliable workarounds. Recent progress in multimodal machine learning, web automation, and other fields is making it possible to build assistive agents that seamlessly integrate with users’ experiences. In line with ICMI 2026’s theme of context and cultural awareness for multimodal interaction, this workshop focuses on assistive agents for accessible computing in real-world interfaces, where language and cultural variation can affect what works in practice. We bring together researchers working on accessibility, multimodal interaction, and human-centered AI for interactive systems. The workshop includes keynotes, panels, paper presentations, posters, and structured breakout discussions. To further accelerate progress, we also host the Navigate-and-Explain Community Project, a unique ACCESS-MI initiative where participants contribute to an assistive agent that will ultimately be deployed and distributed to visually impaired people in the real world.
Organisers:
- Santosh Patapati (Texas, USA)
- Rahul Kumar Mehrotra (Bennett University, Uttar Pradesh, India)
- Trisanth Srinivasan (Texas, USA)
- Sowmya Kirkpatrick (Meta, New York, USA)
- Ashwini Joshi (Warner Bros, Washington, USA)
—
LaugHSMI: Laughter, Humour, Smiles in Multimodal Interactions
Click here to go to the workshop site
LaugHSMI (Laughter, Humour, Smiles in Multimodal Interactions) is a workshop dedicated to advancing research on the role of laughter, smiles, and humor in human-computer interaction and multimodal communication. These phenomena are fundamental aspects of human social interaction, yet they remain challenging to detect, interpret, and generate in computational systems.
Laughter and smiling serve multiple communicative functions beyond expressing amusement—they facilitate social bonding, regulate conversation flow, signal understanding, and convey complex emotional states. Humor adds another layer of complexity, involving cognitive, linguistic, and cultural dimensions. Understanding and modeling these phenomena is crucial for creating more natural, engaging, and socially intelligent interactive systems.
This workshop brings together researchers from affective computing, natural language processing, computer vision, speech processing, human-computer interaction, and social signal processing to address the unique challenges posed by laughter, smiles, and humor in multimodal interactions. We aim to foster interdisciplinary dialogue and advance the state-of-the-art in detecting, analyzing, and generating these important social signals.
Organisers:
- Valentin Barriere (University of Chile)
- Sofia Callejas (Université Paris-Saclay & Universidad de Chile)
- Vladislav Maraev (University of Gothenburg)
- Chiara Mazzocconi (Aix-Marseille Université)
- Catherine Pelachaud (Sorbonne University)
- Brian Ravenet (Université Paris-Saclay)
—
Collective States in Multimodal Interaction
Click here to go to the workshop site
This workshop explores how multimodal sensing and AI techniques can be utilized to detect and interpret the collective states that emerge in group interactions, in conjunction with individual participant states.
Organisers:
- Teruhisa Misu (Honda Research Institute USA, Inc.)
- Zhaobo Zheng (Honda Research Institute USA, Inc.)
- Koji Inoue (Kyoto Universtiy)
- Chikara Maeda (Honda Research Institute Japan)
—
Cross-Cultural Multimodal Interaction (CCMI)
Click here to go to the workshop site
This ICMI 2026 workshop aims to establish an international research platform to explore how linguistic and cultural differences shape nonverbal behavior and interaction dynamics. Building on the success of our first workshop, this second edition shifts the focus from problem identification to concrete action and methodological evaluation. This year’s discussion-centric workshop focuses on two key themes: (1) Concrete Data Collection & Case Studies: Sharing practical “Case Reports” to overcome the logistical hurdles of multi-site data collection, with the goal of drafting a roadmap for truly cross-cultural multimodal datasets. (2) Evaluating MLLMs in Cultural Contexts: Developing methodologies to benchmark Multimodal Large Language Models (MLLMs) for cultural sensitivity, specifically examining their ability to handle cultural nuances in gestures, facial expressions, and turn-taking.
Organisers:
- Koji Inoue (Kyoto University)
- Shogo Okada (Japan Advanced Institute of Science and Technology (JAIST))
- Divesh Lala (Kyoto University)
- Taiga Mori (Kyoto University)
- Sahba Zojaji (The Chinese University of Hong Kong, Shenzhen)
- Nancy F. Chen (Agency for Science, Technology, and Research (A*STAR))
- Yukiko I. Nakano (Seikei University)
- Tatsuya Kawahara (Kyoto University)
—
The Sixth International Workshop on Automated Assessment of Pain (AAP)
Click here to go to the workshop site
Pain typically is measured by patient self-report, but self-reported pain is difficult to interpret and may be impaired or in some circumstances not possible to obtain. For instance, in patients with restricted verbal abilities such as neonates, young children, and in patients with certain neurological or psychiatric impairments (e.g., dementia). Additionally, the subjectively experienced pain may be partly or even completely unrelated to the somatic pathology of tissue damage and other disorders. Therefore, the standard self-assessment of pain does not always allow for an objective and reliable assessment of the quality and intensity of pain. Given individual differences among patients, their families, and healthcare providers, pain often is poorly assessed, underestimated, and inadequately treated. To improve assessment of pain, objective, valid, and efficient assessment of the onset, intensity, and pattern of occurrence of pain is necessary. To address these needs, several efforts have been made in the machine learning and computer vision communities for automatic and objective assessment of pain from video as a powerful alternative to self-reported pain. The workshop aims to bring together interdisciplinary researchers working in the field of automatic multimodal assessment of pain (using video and physiological signals). A key focus of the workshop is the translation of laboratory work into clinical practice.
Organisers:
- Zakia Hammal (The Robotics Institute, Carnegie Mellon University, USA)
- Steffen Walter (University Hospital Ulm, Germany)
- Nadia Berthouze (University College London, UK)
—
HeMAI – Multimodal Interaction with Generative AI Health Applications
Click here to go to the workshop site
This full-day workshop at ICMI 2026 explores the transformative potential of multimodal interaction for generative AI health applications. The workshop will address key research challenges and opportunities across diverse input and output modalities, including language, speech, vision, gesture, physiological signals, and interactive visualizations. With a focus on interactive, collaborative decision-making, the workshop will cover essential topics such as personalized systems, XAI methods for transparency and trust, synthetic data and simulation, and user-centered design principles for clinical and patient-facing scenarios. Experts and practitioners are invited to join a collaborative environment of paper presentations, panel discussions, and keynote speakers to shape the future of multimodal, intelligent systems in healthcare.
Organisers:
- Stefan Hillmann (Technische Universität Berlin, Germany)
- Sebastian Möller (Technische Universität Berlin & DFKI Berlin, Germany)
- Catherine Pelachaud (CNRS – ISIR, Sorbonne University, France)
- Lisa Raithel (Technische Universität Berlin / BIFOLD & DFKI, Germany)
- Roland Roller (DFKI Berlin & Technische Universität Berlin)
—


