出版社:The Japanese Society for Artificial Intelligence
摘要:In this study, the ability estimation in the job interview scene was carried out. In the field of communication between human and human, the communication ability estimation using multimodal information shows high accuracy. It is considered that the same ability estimation is possible in the dialogue between the interviewee and the interviewer. In this study, we developed a model to estimate the evaluation of interviewees by using multimodal information such as speech features, prosodic features, motion features, and head features such as head movement and gaze. In the evaluation of the interviewee, the following were used: Social basic ability determined by the Ministry of Economy, Trade and Industry and JAVADA determined by the Ministry of Health, Labour and Welfare. As a result of the evaluation experiment using SVM, in the evaluation item of "posture", the accuracy of language and action feature set showed 0.89, and in "assertion of opinion", the accuracy of action and head feature set showed 0.87. And, the weight of the feature quantity which contributed to the estimation was examined in order to investigate the relation between each evaluation item and multimodal information. In this paper, from these results, the relation between multimodal information and evaluation in the interview scene is described.