Multimodal Age and Gender Classification Using Ear and Profile Face Images

Yaman D., Eyiokur F. I., Ekenel H. K.

32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), California, United States Of America, 16 - 20 June 2019, pp.2414-2421 identifier identifier

  • Publication Type: Conference Paper / Full Text
  • Volume:
  • Doi Number: 10.1109/cvprw.2019.00296
  • City: California
  • Country: United States Of America
  • Page Numbers: pp.2414-2421
  • Istanbul Technical University Affiliated: Yes


In this paper, we present multimodal deep neural network frameworks for age and gender classification, which take input a profile face image as well as an ear image. Our main objective is to enhance the accuracy of soft biometric trait extraction from profile face images by additionally utilizing a promising biometric modality: ear appearance. For this purpose, we provided end-to-end multimodal deep learning frameworks. We explored different multimodal strategies by employing data, feature, and score level fusion. To increase representation and discrimination capability of the deep neural networks, we benefited from domain adaptation and employed center loss besides softmax loss. We conducted extensive experiments on the UND-E UND-J2, and FERET datasets. Experimental results indicated that profile face images contain a rich source of information for age and gender classification. We found that the presented multimodal system achieves very high age and gender classification accuracies. Moreover, we attained superior results compared to the state-of-the-art profile face image or ear image-based age and gender classification methods.