[visionlist] competitions at FG 2020
dacevedo at dc.uba.ar
Mon Nov 4 16:31:44 -04 2019
*** Please accept our apologies if you receive multiple copies of this
We invite you to participate in the open competitionsat The 15th IEEE
International Conference on Automatic Face and Gesture Recognition (FG
2020) that will be held in Buenos Aires, Argentina, 18-22 May 2020.
2020 ChaLearn LAP Workshop FG: Identity-preserving human detection
EmoPain Face and Movement Behaviour Challenge
4th Recognizing Families In the Wild (RFIW)
Joint Challenge on Compound Emotion Recognition and Multimodal
(Audio, Facial and Gesture) based Emotion Recognition (CER\&MMER)
Affective Behavior Analysis in-the-wild
Check for more information at https://fg2020.org/competitions/
** 2020 ChaLearn LAP Workshop FG: Identity-preserving human detection
Participants will be asked to automatically produce human detections on
depth and/or radiometric thermal images ïn the form of 2D bounding box
locations. These two visual modalities preserve subjects’ identity to a
large extent while still providing rich human-related features: shape
and absolute temperature. We provide a dataset consisting of more than
100K multimodal frames, which are a mix of close-range in-the-wild
pedestrian scenes and indoor ones with people performing basic actions
in scripted scenarios. The competition is divided into 3 tracks, so the
participants can choose to participate in one or more and will be asked
to exploit different input data for the task: (1) only depth, (2) only
thermal, and (3) both depth and thermal (spatially registered to depth).
Competition website: http://chalearnlap.cvc.uab.es/challenge/34/description/
Competition start-ending: November 26th, 2019 — February 4th, 2020
- Albert Clapés, Computer Vision Center (Universitat Autònoma de
Barcelona), aclapes at cvc.uab.es
- Carla Morral, Universitat de Barcelona, carla.morral at gmail.com
- Julio C. S. Jacques Junior, Universitat Oberta de Catalunya & Computer
Vision Center (Universitat Autònoma de Barcelona), juliojj at gmail.com
- Sergio Escalera, Universitat de Barcelona & Computer Vision Center
(Universitat Autònoma de Barcelona), sergio at maia.ub.es
** EmoPain Face and Movement Behaviour Challenge **
The EmoPain Face and Movement Behaviour Challenge is based on the
EmoPain dataset captured from chronic pain patients and healthy
participants performing movement exercises. The challenge consists of
three tasks to choose from: ‘Pain Intensity Estimation from Facial
Expressions’, ‘Pain Level Recognition from Multimodal Movement Data’,
‘Multimodal Movement Behaviour Classification’. The challenge is an
opportunity to contribute to solving the challenging problem of
automatic detection of pain behaviours and pain levels during movement
performance. This is fundamental to the development of technology that
improves the quality and quantity of engagement in valued everyday
activities for people with chronic pain by providing tailored support to
the specific barriers that arise.
Competition website: https://mvrjustid.github.io/EmoPainChallenge2020/
Competition start-ending dates: October 2019 – Jan 17, 2020
- Prof Nadia Berthouze, UCL
- Dr. Amanda Williams, UCL,
- Dr Michel Valstar, University of Nottingham,
- Dr Hongying Meng, Brunel University London,
- Dr Min Aung, University of East Anglia,
- Dr Nicholas Lane, University of Oxford.
- Dr Joy Egede, University of Nottingham,
- Dr Olugbade Temitayo, UCL,
- Wang Chongyang , UCL,
- Siyang Song, University of Nottingham.
** 4th Recognizing Families In the Wild (RFIW) **
We are pleased to kick-off the 4th RFIW data challenge workshop in
conjunction with 2020 IEEE AMFG conference.
New this year:
Support 3 tasks (2 new)– upon successful Kaggle competition, where many
obtained impressive verification results, we see it as time for newer,
more practical challenges (i.e. large-scale tri-subject verification and
search & retrieval for relatives of missing children).
Call for novel work in automatic kinship recognition (i.e. general paper
Brave New Ideas: new ways of viewing problem and its formulation.
Special attention for inter-disciplinary work and innovative use-cases.
Major release of FIW (v1.1.0)
Submissions will be peer-reviewed for the proceedings of 4th RFIW
workshop at 2020 AMFG– papers will be accepted as orals and posters
Competition website: https://web.northeastern.edu/smilelab/rfiw2020/
Competition start-ending dates: October 18, 2019
Joseph Robinson robinson.jo at husky.neu.edu. Department of Electrical and
Computer Engineering, Northeastern University, Boston, MA, USA
** Joint Challenge on Compound Emotion Recognition and Multimodal
(Audio, Facial and Gesture) based Emotion Recognition (CER&MMER) **
Emotion recognition has a key role in affective computing. People
express emotions through different modalities. Expanding the focus to
several expression forms can facilitate research on emotion recognition
as well as human-machine interaction. This competition focuses on two
important problems which are: (1) recognition of compound emotions, that
require, in addition to performing an effective visual analysis, to deal
with recognition of micro emotions. The database includes 31250 facial
faces with different emotions of 115 subjects; (2) recognition of
multimodal emotions composed of three modalities, namely, facial
expressions, body movement and gestures, and speech.
- 5th Nov, 2019: Beginning of the quantitative competition, release of
development and data.
- 10th Feb, 2020: Deadline for code submission.
- 22nd Feb, 2020: Release of final evaluation data decryption key.
Participants start predicting the results on the final evaluation data.
- 23rd Feb, 2020: Contest paper submission deadline.
Check for more deadlines at
- Gholamreza Anbarjafari
- Sergio Escalera
- Kamal Nasrollahi
- Thomas B. Moeslund
- Dorota Kiaminska
- Cagri Ozcinar
- Tomasz Spanski
- Juri Allik
** Affective Behavior Analysis in-the-wild **
Representing human emotions has been a basic topic of research. The most
usual emotion representation (ER) is the categorical one (7 basic
categories). Discrete ER can also be described in terms of the FACS
model, in which all possible facial actions are described in terms of
Action Units (AUs). Finally, the dimensional model of affect -with
valence and arousal (VA) being the most usual representations- has been
proposed to distinguish between subtly different displays of affect and
encode small changes in the intensity of each emotion on a continuous
scale. This Competition is split into 3 Challenges: i) VA estimation,
ii) 7 basic emotion classification, iii) AU detection. These Challenges
use the Aff-Wild2, the first comprehensive benchmark for all 3 affect
recognition tasks in-the-wild
Competition start-ending dates:
- 01 Nov 2019: Call for participation announced; start of Competition
- 14 Feb 2020: Results and Paper submission deadline
- 21 Feb 2020: End of review; Decisions sent to authors; Winner announcement
- 28 Feb 2020: Camera ready versions deadline
- 5 Mar 2020: End of Competition
- Stefanos Zafeiriou, Imperial College London, UK
(s.zafeiriou at imperial.ac.uk)
- Dimitrios Kollias, Imperial College London, UK
(dimitrios.kollias15 at imperial.ac.uk)
- Attila Schulc, Realeyes (attila.schulc at realeyesit.com)
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the visionlist