Call for Papers: HCI International 2024 Session

26th International Conference on Human-Computer Interaction

29 June – 04 July 2024

Session

“Semantic, artificial and computational interaction studies: Towards a behavioromics of multimodal communication”

Call for papers

A common theme that has emerged across different disciplines, especially in recent years, is the transition from monomodal to multimodal domains. This includes, for example, multimodal learning – learning representations from multiple data modalities – , multimodal communication – the study of gestures, facial expressions, head movements, and the like in addition to speech – , or multimodal logic – that is, modal logic with more than one primitive operator. While the predominant use of “multimodal communication” (and its relatives) refers to human-human interaction, other domains can also be conceived in terms of interacting modes. The field of HCI and HRI, with its focus on natural (for humans) interfaces, has paid attention to multimodality from an early stage. It is also an important topic in conversation analysis and cognitive science and is beginning to percolate into information science and theoretical linguistics. At the same time, due to the digital turn, work on multimodality is being expanded by data analytics using machine learning tools for modelling, detecting, analysing and simulating any form of communication. Relatedly, but independently, symbol grounding approaches have been multimodally extended, and advances in computer vision and multimodal AI are prolific. However, while these fields share a common empirical domain, there is little interaction between them. This session aims to bring these branches together – a collaborative endeavour that we call “behaviouromics”.

The session is open, but not restricted, to topics such as the following:

  • multimodal generative AI
  • predictive multimodal computing
  • dialogue generation, dialogue systems and dialogue semantics
  • social robot interaction and adaptive behaviour
  • monitoring and processing in interaction
  • (big) multimodal data
  • multimodal data analytics
  • automatic multimodal annotation beyond written text
  • virtual reality and augmented reality applications
  • simulation-based learning in virtual environments
  • representation schemes for multimodal communication
  • verbal and non-verbal social signalling in humans and non-humans
  • the role of multimodality in 4E cognition
  • notions and theories of multimodality
  • multimodality in logics

We want to emphasize that conceptual contributions are highly welcome!

The conference session aims to provide a platform to bring together computer scientists, linguists, psychologists and researchers in related fields who are working on multimodal interaction. We are all working on almost the same topic from different angles, but there are far too few opportunities to interact. But sharing and seeing what others are doing is crucial for the methodological, empirical and theoretical challenges outlined above. The planned session will support this collaboration.

The conference session will take place *virtually* in conjunction with HCI International 2024.

Full papers will be published as part of the conference proceedings by Springer.

If you want to contribute, please send a message to any one of the organizers until 26 October 2023:
Alexander Mehler (mehler@em.uni-frankfurt.de)
Andy Lücking (luecking@em.uni-frankfurt.de)
Alexander Henlein (henlein@em.uni-frankfurt.de)

Important dates

until 26 October 2023send email message to one of the session organizers
30 October 2023upload abstract (up to 500 words)
02 February 2024full paper is due
29 June – 04 July 2024HCI International conference (virtual)

Session organizers
Alexander Mehler
Andy Lücking
Alexander Henlein

Leave a Reply