How Do You Feel Online
Citations Over TimeTop 20% of 2020 papers
Abstract
Emotions are an intrinsic part of the social media user experience that can evoke negative behaviors such as cyberbullying and trolling. Detecting the emotions of social media users may enable responding to and mitigating these problems. Prior work suggests this may be achievable on smartphones: emotions can be detected via built-in sensors during prolonged input tasks. We extend these ideas to a social media context featuring sparse input interleaved with more passive browsing and media consumption activities. To achieve this, we present two studies. In the first, we elicit participant's emotions using images and videos and capture sensor data from a mobile device, including data from a novel passive sensor: its built-in eye-tracker. Using this data, we construct machine learning models that predict self-reported binary affect, achieving 93.20% peak accuracy. A follow-up study extends these results to a more ecologically valid scenario in which participants browse their social media feeds. The study yields high accuracies for both self-reported binary valence (94.16%) and arousal (92.28%). We present a discussion of the sensors, features and study design choices that contribute to this high performance and that future designers and researchers can use to create effective and accurate smartphone-based affect detection systems.
Related Papers
- → Effects of Emotional Valence and Arousal on Time Perception(2014)35 cited
- → What are the influences of orthogonally-manipulated valence and arousal on performance monitoring processes? The effects of affective state(2013)38 cited
- → Emotion Recognition from Physiological Signals Using Parallel Stacked Autoencoders(2018)22 cited
- Exploring the Effect of Arousal and Valence on Mouse Interaction(2013)
- → Valence evaluation with approaching or withdrawing cues: directly testing valence–arousal conflict theory(2017)5 cited