Multimodal first impression analysis with deep residual networks
until further notice
Number of pages
SourceIEEE Transactions on Affective Computing, 9, 3, (2018), pp. 316-329
Article / Letter to editor
Display more detailsDisplay less details
SW OZ DCC CO
SW OZ DCC KI
IEEE Transactions on Affective Computing
SubjectAction, intention, and motor control; Cognitive artificial intelligence; DI-BCB_DCC_Theme 2: Perception, Action and Control; DI-BCB_DCC_Theme 4: Brain Networks and Neuronal Communication
People form first impressions about the personalities of unfamiliar individuals even after very brief interactions with them. In this study we present and evaluate several models that mimic this automatic social behavior. Specifically, we present several models trained on a large dataset of short YouTube video blog posts for predicting apparent Big Five personality traits of people and whether they seem suitable to be recommended to a job interview. Along with presenting our audiovisual approach and results that won the third place in the ChaLearn First Impressions Challenge, we investigate modeling in different modalities including audio only, visual only, language only, audiovisual, and combination of audiovisual and language. Our results demonstrate that the best performance could be obtained using a fusion of all data modalities.
Upload full text
Use your RU credentials (u/z-number and password) to log in with SURFconext to upload a file for processing by the repository team.