Please use this identifier to cite or link to this item: http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/1226
Full metadata record
DC FieldValueLanguage
dc.contributor.authorJyoti, S.
dc.contributor.authorDhall, A.
dc.date.accessioned2019-05-14T13:39:15Z
dc.date.available2019-05-14T13:39:15Z
dc.date.issued2019-05-14
dc.identifier.urihttp://localhost:8080/xmlui/handle/123456789/1226
dc.description.abstractEye gaze estimation is an important problem in automatic human behavior understanding. This paper proposes a deep learning based method for inferring the eye gaze direction. The method is based on the use of ensemble of networks, which capture both the geometric and texture information. Firstly, a Deep Neural Network (DNN) is trained using the geometric features that are extracted from the facial landmark locations. Secondly, for the texture based features, three Convolutional Neural Networks (CNN) are trained i.e. for the patch around the left eye, right eye, and the combined eyes, respectively. Finally, the information from the four channels is fused with concatenation and dense layers are trained to predict the final eye gaze. The experiments are performed on the two publicly available datasets: Columbia eye gaze and TabletGaze. The extensive evaluation shows the superior performance of the proposed framework. We also evaluate the performance of the recently proposed swish activation function as compared to Rectified Linear Unit (ReLU) for eye gaze estimation.en_US
dc.language.isoen_USen_US
dc.titleAutomatic eye gaze estimation using geometric and texture-based networksen_US
dc.typeArticleen_US
Appears in Collections:Year-2018

Files in This Item:
File Description SizeFormat 
Full Text.pdf1.59 MBAdobe PDFView/Open    Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.