The chair of EIHW (co-)organise a variety of events and challenges, details are posted bellow. 


Tag der Informatik

Look out for two presentations by members of EIHW during the informatiks day. 


Challenges & Workshops

ICMI 2018 Eating Analysis & Tracking Challenge 


The ICMI 2018 Eating Analysis & Tracking Challenge is an open research competition dealing with Machine Learning for audio/visual tracking of human subjects’ recorded while eating different types of food during speaking.

The Challenge features three Sub-Challenges:

The Audio/Visual Emotion Challenge and Workshop (AVEC 2018)

The Audio/Visual Emotion Challenge and Workshop (AVEC 2018) Bipolar Disorder and Cross-cultural Affect” is a satellite event of ACM MM 2018 and the eighth competition aimed at comparison of multimedia processing and machine learning methods for automatic audio, visual, and audio-visual health and emotion sensing, with all participants competing under strictly the same conditions.

The goal of the Challenge is to provide a common benchmark test set for multimodal information processing and to bring together the audio, visual and audio-visual affect recognition communities, to compare the relative merits of the approaches to automatic health and emotion analysis under well-defined conditions. Another motivation is the need to advance health and emotion recognition systems to be able to deal with fully naturalistic behaviour in large volumes of un-segmented, non-prototypical and non-preselected data, as this is exactly the type of data that both multimedia and human-machine/human-robot communication interfaces have to face in the real world.

We are calling for teams to participate in three Sub-Challenges:

The Affective Social Multimedia Computing 2018

An ACMMM 2018 Setellite Workshop 

Affective social multimedia computing is an emergent research topic for both affective computing and multimedia research communities. Social multimedia is fundamentally changing how we communicate, interact, and collaborate with other people in our daily lives. Comparing with well-organized broadcast news and professionally made videos such as commercials, TV shows, and movies, social multimedia computing imposes great challenges to research communities. Social multimedia contains much affective information. Effective extraction of affective information from social multimedia can greatly help social multimedia computing (e.g., processing, index, retrieval, and understanding). Although much progress has been made in traditional multimedia research on multimedia content analysis, indexing, and retrieval based on subjective concepts such as emotion, aesthetics, and preference, affective social multimedia computing is a new research area. The affective social multimedia computing aims to proceed affective information from social multi-media. For massive and heterogeneous social media data, the research requires multidisciplinary understanding of content and perceptional cues from social multimedia. From the multimedia perspective, the research relies on the theoretical and technological findings in affective computing, machine learning, pattern recognition, signal/multimedia processing, computer vision, speech processing, behavior and social psychology. Affective analysis of social multimedia is attracting growing attention from industry and businesses that provide social networking sites, content-sharing services, distribute and host the media. This workshop focuses on the analysis of affective signals in social multimedia (e.g., twitter, weichat, weibo, youtube, facebook, etc).

The workshop will address, but is not limited to, the following topics:


Interspeech ComParE 2018


The Interspeech 2018 Computational Paralinguistics ChallengE (ComParE) is an open Challenge dealing with states and traits of speakers as manifested in their speech signal’s acoustic properties. There have so far been nine consecutive Challenges at INTERSPEECH since 2009 (cf. the Challenge series‘ repository at, but there still exists a multiplicity of not yet covered, but highly relevant paralinguistic phenomena. Thus, in this year’s 10th anniversary edition, we introduce four new tasks.

The following Sub-Challenges are addressed:

  • Atypical Affect Sub-Challenge, emotion of disabled speakers is to be recognised.
  • Self-Assessed Affect Sub-Challenge, self-assessed affect shall be determined.
  • Crying Sub-Challenge, mood-related types of infant vocalisation have to be classified.
  • Heart Beats Sub-Challenge, types of Heart Beat Sounds need to be distinguished.