Emoti W 2016: Video and group-level emotion recognition challenges

Abhinav Dhall, Roland Goecke, Jyoti Joshi, Jesse Hoey, Tom Gedeon

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    139 Citations (Scopus)

    Abstract

    This paper discusses the baseline for the Emotion Recognition in the Wild (EmotiW) 2016 challenge. Continuing on the theme of automatic affect recognition in the wild', the Emoti W challenge 2016 consists of two sub-challenges: An audio-video based emotion and a new group-based emotion recognition sub-challenges. The audio-video based subchallenge is based on the Acted Facial Expressions in the Wild (AFEW) database. The group-based emotion recognition sub-challenge is based on the Happy People Images (HAPPEI) database. We describe the data, baseline method, challenge protocols and the challenge results. A total of 22 and 7 teams participated in the audio-video based emotion and group-based emotion sub-challenges, respectively.

    Original languageEnglish
    Title of host publicationICMI 2016 - Proceedings of the 18th ACM International Conference on Multimodal Interaction
    EditorsCatherine Pelachaud, Yukiko I. Nakano, Toyoaki Nishida, Carlos Busso, Louis-Philippe Morency, Elisabeth Andre
    PublisherAssociation for Computing Machinery (ACM)
    Pages427-432
    Number of pages6
    ISBN (Electronic)9781450345569
    DOIs
    Publication statusPublished - 31 Oct 2016
    Event18th ACM International Conference on Multimodal Interaction, ICMI 2016 - Tokyo, Japan
    Duration: 12 Nov 201616 Nov 2016

    Publication series

    NameICMI 2016 - Proceedings of the 18th ACM International Conference on Multimodal Interaction

    Conference

    Conference18th ACM International Conference on Multimodal Interaction, ICMI 2016
    Country/TerritoryJapan
    CityTokyo
    Period12/11/1616/11/16

    Fingerprint

    Dive into the research topics of 'Emoti W 2016: Video and group-level emotion recognition challenges'. Together they form a unique fingerprint.

    Cite this