dc.contributor.author |
Dhall, A. |
|
dc.contributor.author |
Joshi, J. |
|
dc.contributor.author |
Goecke, R. |
|
dc.contributor.author |
Hoey, J. |
|
dc.contributor.author |
Ghosh, S. |
|
dc.contributor.author |
Gedeon, T. |
|
dc.date.accessioned |
2021-10-06T17:25:30Z |
|
dc.date.available |
2021-10-06T17:25:30Z |
|
dc.date.issued |
2021-10-06 |
|
dc.identifier.uri |
http://localhost:8080/xmlui/handle/123456789/2905 |
|
dc.description.abstract |
Research in automatic affect recognition has come a long way. This
paper describes the fifth Emotion Recognition in the Wild (EmotiW)
challenge 2017. EmotiW aims at providing a common benchmarking
platform for researchers working on different aspects of affective
computing. This year there are two sub-challenges: a) Audio-video
emotion recognition and b) group-level emotion recognition. These
challenges are based on the acted facial expressions in the wild
and group affect databases, respectively. The particular focus of
the challenge is to evaluate method in ‘in the wild’ settings. ‘In the
wild’ here is used to describe the various environments represented
in the images and videos, which represent real-world (not lab like)
scenarios. The baseline, data, protocol of the two challenges and
the challenge participation are discussed in detail in this paper. |
en_US |
dc.language.iso |
en_US |
en_US |
dc.subject |
Audio-video data corpus |
en_US |
dc.subject |
Emotion recognition |
en_US |
dc.subject |
Group-level emotion recognition |
en_US |
dc.subject |
Facial expression challenge |
en_US |
dc.subject |
Affect analysis in the wild |
en_US |
dc.title |
From individual to group-level emotion recognition: emoti W 5.0 |
en_US |
dc.type |
Article |
en_US |