User Profiling Based on Nonlinguistic Audio Data

User profiling refers to inferring people's attributes of interest (AoIs) like gender and occupation, which enables various applications ranging from personalized services to collective analyses. Massive nonlinguistic audio data brings a novel opportunity for user profiling due to the prevalenc...

Full description

Bibliographic Details
Main Authors: Shen, Jiaxing (Author), Cao, Jiannong (Author), Lederman, Oren (Author), Tang, Shaojie (Author), Pentland, Alex (Author)
Format: Article
Language:English
Published: ACM Transactions on Information Systems, 2021-12-07T22:03:01Z.
Subjects:
Online Access:Get fulltext
LEADER 01976 am a22001813u 4500
001 138373
042 |a dc 
100 1 0 |a Shen, Jiaxing  |e author 
700 1 0 |a Cao, Jiannong  |e author 
700 1 0 |a Lederman, Oren  |e author 
700 1 0 |a Tang, Shaojie  |e author 
700 1 0 |a Pentland, Alex  |e author 
245 0 0 |a User Profiling Based on Nonlinguistic Audio Data 
260 |b ACM Transactions on Information Systems,   |c 2021-12-07T22:03:01Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/138373 
520 |a User profiling refers to inferring people's attributes of interest (AoIs) like gender and occupation, which enables various applications ranging from personalized services to collective analyses. Massive nonlinguistic audio data brings a novel opportunity for user profiling due to the prevalence of studying spontaneous face-to-face communication. Nonlinguistic audio is coarse-grained audio data without linguistic content. It is collected due to privacy concerns in private situations like doctor-patient dialogues. The opportunity facilitates optimized organizational management and personalized healthcare, especially for chronic diseases. In this article, we are the first to build a user profiling system to infer gender and personality based on nonlinguistic audio. Instead of linguistic or acoustic features that are unable to extract, we focus on conversational features that could reflect AoIs. We firstly develop an adaptive voice activity detection algorithm that could address individual differences in voice and false-positive voice activities caused by people nearby. Secondly, we propose a gender-assisted multi-task learning method to combat dynamics in human behavior by integrating gender differences and the correlation of personality traits. According to the experimental evaluation of 100 people in 273 meetings, we achieved 0.759 and 0.652 in F1-score for gender identification and personality recognition, respectively. 
546 |a en_US 
655 7 |a Article