The Computational Behavior Lab focuses broadly on multi-modal methods for computational behavior science, specifically in areas of modelling, analysis, and synthesis of human behavior and emotion using diverse sensors.. |
|
|
Dense Body PoseLow-resolution 3D human shape and pose estimation is a challenging problem. We propose a resolution-aware neural network which can deal with different resolution images with a single model. For training the network, we propose a directional self-supervision loss which can exploit the output consistency across different resolutions to remedy the issue of lacking high-quality 3D labels. In addition, we introduce a contrastive feature loss which is more effective than MSE for measuring high-dimensional vectors and helps learn better feature representations. Project: Low-resolution dense pose estimation |
Cognitive Assistant for the Visually ImpairedWe developed a prototype mobile vision system for the visually impaired that performs both person and emotion recognition in diverse environments. Project: ZFace TED Talk: How New Technology Helps Blind People Explore the World |
Automated Facial Action Unit CodingThis study addressed how design choices influence performance in facial AU coding using deep learning systems, by evaluating the combinations of different components and their parameters present in such systems. |
Facial Expression SynthesisThis study proposed a generative approach that achieves 3D geometry based AU manipulation with idiosyncratic loss to synthesize facial expressions. With the semantic resampling, this approach provides a balanced distribution of AU intensity labels, which is crucial to train AU intensity estimators. We have shown that using the balanced synthetic set for training performs better than using the real training dataset on the same test set. The method generalizes to non-frontal views and to unseen domains. |
Smartphone-based physiology measurements
|