Please use this identifier to cite or link to this item: http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/1246
Full metadata record
DC FieldValueLanguage
dc.contributor.authorGhosh, S.-
dc.contributor.authorDhall, A.-
dc.contributor.authorSebe, N.-
dc.date.accessioned2019-05-16T12:42:27Z-
dc.date.available2019-05-16T12:42:27Z-
dc.date.issued2019-05-16-
dc.identifier.urihttp://localhost:8080/xmlui/handle/123456789/1246-
dc.description.abstractThis paper proposes a pipeline for automatic group-level affect analysis. A deep neural network-based approach, which leverages on the facial-expression information, scene information and a high-level facial visual attribute information is proposed. A capsule network-based architecture is used to predict the facial expression. Transfer learning is used on Inception-V3 to extract global image-based features which contain scene information. Another network is trained for inferring the facial attributes of the group members. Further, these attributes are pooled at a group-level to train a network for inferring the group-level affect. The facial attribute prediction network, although is simple yet, is effective and generates result comparable to the state-of-the-art methods. Later, model integration is performed from the three channels. The experiments show the effectiveness of the proposed techniques on three ‘in the wild’ databases: Group Affect Database, HAPPEI and UCLA-Protest database.en_US
dc.language.isoen_USen_US
dc.subjectGroup level affect recognition.en_US
dc.titleAutomatic group affect analysis in images via visual attribute and feature networksen_US
dc.typeArticleen_US
Appears in Collections:Year-2018

Files in This Item:
File Description SizeFormat 
Full Text.pdf300.53 kBAdobe PDFView/Open    Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.