[visionlist] The OMG-Emotion Recognition Challenge

Pablo Barros barros at informatik.uni-hamburg.de
Wed Mar 14 11:34:53 -05 2018


CALL FOR PARTICIPATION

The One-Minute Gradual-Emotion Recognition (OMG-Emotion)
held in partnership with the WCCI/IJCNN 2018 in Rio de Janeiro, Brazil.

https://www2.informatik.uni-hamburg.de/wtm/OMG-EmotionChallenge/


 I. Aim and Scope

Our One-Minute-Gradual Emotion Dataset (OMG-Emotion Dataset) is composed
of 420 relatively long emotion videos with an average length of 1
minute, collected from a variety of Youtube channels. The videos were
selected automatically based on specific search terms related to the
term ``monologue''. Using monologue videos allowed for different
emotional behaviors to be presented in one context and that changes
gradually over time. Videos were separated into clips based on
utterances, and each utterance was annotated by at least five
independent subjects using the Amazon Mechanical Turk tool. To maintain
the contextual information for each video, each annotator watched the
clips of a video in sequence and had to annotate each video using an
arousal/valence scale and a categorical emotion based on the universal
emotions from Ekman.

The participants are encouraged to use crossmodal information in their
models, as the videos were labeled by humans without distinction of any
modality.

II. How to Participate

To participate, please send us an email to
barros at informatik.uni-hamburg.de with the title "OMG-Emotion Recognition
Team Registration". This e-mail must contain the following information:
Team Name
Team Members
Affiliation

Each team can have a maximum of 5 participants. You will receive from us
the access to the dataset and all the important information about how to
train and evaluate your models.
For the final submission, each team will have to send us a .csv file
containing the final arousal/valence values for each of the utterances
on the test dataset. We also request a link to a GitHub repository where
your solution must be stored, and a link to an ArXiv paper with 4-6
pages describing your model and results. The best papers will be invited
to submit their detailed research to a journal yet to be specified.
Also, the best participating teams will hold an oral presentation about
their solution during the WCCI/IJCNN 2018 conference.


III. Important Dates

Publishing of training and validation data with annotations: March 14,
2018.
Publishing of the test data, and an opening of the online submission:
April 11, 2018.
Closing of the submission portal: April 13, 2018.
Announcement of the winner through the submission portal: April 18, 2018.



IV. Organization

Pablo Barros, University of Hamburg, Germany
Egor Lakomkin, University of Hamburg, Germany
Henrique Siqueira, Hamburg University, Germany
Alexander Sutherland, Hamburg University, Germany
Stefan Wermter, Hamburg University, Germany

--------------------------

-- 
Dr.rer.nat. Pablo Barros
Postdoctoral Research Associate - Crossmodal Learning Project (CML)
Knowledge Technology
Department of Informatics
University of Hamburg
Vogt-Koelln-Str. 30
22527 Hamburg, Germany
Phone: +49 40 42883 2535
Fax: +49 40 42883 2515
barros at informatik.uni-hamburg.de
https://www.inf.uni-hamburg.de/en/inst/ab/wtm/people/barros.html
https://www.inf.uni-hamburg.de/en/inst/ab/wtm/




More information about the visionlist mailing list