Call for Papers: The Second Workshop on Multimodal Semantic Representations (MMSR II)

KL
Kenneth Lai
Fri, Apr 12, 2024 5:00 PM

CALL FOR PAPERS

The Second Workshop on Multimodal Semantic Representations (MMSR II)
Co-located with ECAI 2024 (https://www.ecai2024.eu/)
19-24 October, Santiago de Compostela, Spain
(workshop on 19 or 20 October)

Workshop website: https://mmsr-workshop.github.io/

Description
The demand for more sophisticated natural human-computer and human-robot
interactions is rapidly increasing as users become more accustomed to
conversation-like interactions with AI and NLP systems. Such interactions
require not only the robust recognition and generation of expressions
through multiple modalities (language, gesture, vision, action, etc.), but
also the encoding of situated meaning.

When communications become multimodal, each modality in operation provides
an orthogonal angle through which to probe the computational model of the
other modalities, including the behaviors and communicative capabilities
afforded by each. Multimodal interactions thus require a unified framework
and control language through which systems interpret inputs and behaviors
and generate informative outputs. This is vital for intelligent and often
embodied systems to understand the situation and context that they inhabit,
whether in the real world or in a mixed-reality environment shared with
humans.

Furthermore, multimodal large language models appear to offer the
possibility for more dynamic and contextually rich interactions across
various modalities, including facial expressions, gestures, actions, and
language. We invite discussion on how representations and pipelines can
potentially integrate such state-of-the-art language models.

We solicit papers on multimodal semantic representation, including but not
limited to the following topics:

  • Semantic frameworks for individual linguistic co-modalities (e.g.
    gaze, facial expression);
  • Formal representation of situated conversation and embodiment,
    including knowledge graphs, designed to represent epistemic state;
  • Design, annotation, and corpora of multimodal interaction and meaning
    representation;
  • Challenges (including cross-lingual and cross-cultural) in multimodal
    representation and/or processing;
  • Criteria or frameworks for evaluation of multimodal semantics;
  • Challenges in aligning co-modalities in formal representation and/or
    NLP tasks;
  • Design and implementation of neurosymbolic or fusion models for
    multimodal processing (with a representational component);
  • Methods for probing knowledge of multimodal (language and vision)
    models;
  • Virtual and situated agents that embody multimodal representations of
    common ground.

Submission Information
Two types of submissions are solicited: long papers and short papers. Long
papers should describe original research and must not exceed 8 pages,
excluding references. Short papers (typically system or project
descriptions, or ongoing research) must not exceed 4 pages, excluding
references. Both types will be published in the workshop proceedings.
Accepted papers get an extra page in the camera-ready version.

We strongly encourage students to submit to the workshop.

Important Dates
May 15, 2024: Submissions due
July 1, 2024: Notification of acceptance decisions
August 2, 2024: Camera-ready papers due

Papers should be formatted using the ECAI style files, available at:
https://www.ecai2024.eu/calls/main-track

Papers will be submitted in PDF format via Chairing Tool at the following
link: https://chairingtool.com/conferences/2MMSR24/MainTrack

Please do not hesitate to reach out with any questions.

Best regards,
Richard Brutti, Lucia Donatelli, Nikhil Krishnaswamy, Kenneth Lai, & James
Pustejovsky
MMSR II organizers
Web page: https://mmsr-workshop.github.io/

CALL FOR PAPERS *The Second Workshop on Multimodal Semantic Representations (MMSR II)* Co-located with ECAI 2024 (https://www.ecai2024.eu/) 19-24 October, Santiago de Compostela, Spain (workshop on 19 or 20 October) *Workshop website*: https://mmsr-workshop.github.io/ *Description* The demand for more sophisticated natural human-computer and human-robot interactions is rapidly increasing as users become more accustomed to conversation-like interactions with AI and NLP systems. Such interactions require not only the robust recognition and generation of expressions through multiple modalities (language, gesture, vision, action, etc.), but also the encoding of situated meaning. When communications become multimodal, each modality in operation provides an orthogonal angle through which to probe the computational model of the other modalities, including the behaviors and communicative capabilities afforded by each. Multimodal interactions thus require a unified framework and control language through which systems interpret inputs and behaviors and generate informative outputs. This is vital for intelligent and often embodied systems to understand the situation and context that they inhabit, whether in the real world or in a mixed-reality environment shared with humans. Furthermore, multimodal large language models appear to offer the possibility for more dynamic and contextually rich interactions across various modalities, including facial expressions, gestures, actions, and language. We invite discussion on how representations and pipelines can potentially integrate such state-of-the-art language models. We solicit papers on multimodal semantic representation, including but not limited to the following topics: - Semantic frameworks for individual linguistic co-modalities (e.g. gaze, facial expression); - Formal representation of situated conversation and embodiment, including knowledge graphs, designed to represent epistemic state; - Design, annotation, and corpora of multimodal interaction and meaning representation; - Challenges (including cross-lingual and cross-cultural) in multimodal representation and/or processing; - Criteria or frameworks for evaluation of multimodal semantics; - Challenges in aligning co-modalities in formal representation and/or NLP tasks; - Design and implementation of neurosymbolic or fusion models for multimodal processing (with a representational component); - Methods for probing knowledge of multimodal (language and vision) models; - Virtual and situated agents that embody multimodal representations of common ground. *Submission Information* Two types of submissions are solicited: long papers and short papers. Long papers should describe original research and must not exceed 8 pages, excluding references. Short papers (typically system or project descriptions, or ongoing research) must not exceed 4 pages, excluding references. Both types will be published in the workshop proceedings. Accepted papers get an extra page in the camera-ready version. We strongly encourage students to submit to the workshop. *Important Dates* May 15, 2024: Submissions due July 1, 2024: Notification of acceptance decisions August 2, 2024: Camera-ready papers due Papers should be formatted using the ECAI style files, available at: https://www.ecai2024.eu/calls/main-track Papers will be submitted in PDF format via Chairing Tool at the following link: https://chairingtool.com/conferences/2MMSR24/MainTrack Please do not hesitate to reach out with any questions. Best regards, Richard Brutti, Lucia Donatelli, Nikhil Krishnaswamy, Kenneth Lai, & James Pustejovsky MMSR II organizers Web page: https://mmsr-workshop.github.io/