MultiMediate: Multi-modal Behaviour Analysis for Artificial Mediation

Datasets

This year's multi-domain engagement task makes use of the NOXI, NOXI-J, PInSoRo and MPIIGroupInteraction datasets. Pre-computed features can be downloaded for all datasets using the links below. For NOXI, NOXI-J, and MPIIGroupInteraction we also publish the original videos. For PInSoRo publishing videos is not possible due to privacy requirements.

MPIIGroupInteraction (MultiMediate'21-'26)

Needed for:

  • Bodily behaviour recognition
  • Backchannel detection
  • Agreement estimation from backchannels
  • Eye contact detection
  • Next speaker prediction
  • Multi-Domain Engagement Estimation

NoXi (MultiMediate'23-'26)

Training set for Multi-Domain Engagement Estimation

NoXi+J (MultiMediate'25-'26)

Japanese and Chinese training set for Multi-Domain Engagement Estimation

PInSoRo (MultiMediate'26)

English-language child-child and child-robot free-play interactions