Integrating Theory-Driven and Data-Driven Approaches to Affective Computing via Deep Probabilistic Programming

Preamble:

This tutorial was taught at Affective Computing and Intelligent Interaction 2019, in September 2019 in Cambridge, UK. This tutorial is in turn based off material introduced in the following paper:

Ong, D. C., Soh, H., Zaki, J., & Goodman, N. D. (in press). Applying Probabilistic Programming to Affective Computing. IEEE Transactions on Affective Computing
[ arXiv ]

Tutorial Details.

  • Location: Computer Laboratory, University of Cambridge. William Gates Building, 15 JJ Thomson Ave.
  • Date and Time: Tuesday 3 September 2019, 9am-12pm.
  • Download the slides presented during the tutorial session.

Presenter: Desmond C. Ong

Contributors:

  • Desmond C. Ong, National University of Singapore (NUS) & Agency for Science, Technology and Research (A*STAR)
  • Zhi-Xuan Tan, Agency for Science, Technology and Research (A*STAR); now at Massachusetts Institute of Technology (MIT)
  • Harold Soh, National University of Singapore (NUS)
  • Jamil Zaki, Stanford University
  • Noah D. Goodman, Stanford University

Abstract

Research in affective computing has traditionally fallen into either theory-driven approaches that may not scale well to the complexities of naturalistic data, or atheoretic, data-driven approaches that learn to recognize complex patterns but still fall short of reasoning about emotions. In this tutorial, we introduce deep probabilistic programming, a new paradigm that models psychologically-grounded theories of emotion using stochastic programs. Specifically, this flexible framework combines the benefits of probabilistic models with those of deep learning models, marrying the advantages of both approaches. For example, when modelling someone’s emotions in context, we may choose to compose a probabilistic model of emotional appraisal with a deep learning model for recognizing emotions from faces—and we can do this all within a single unified framework for training and performing inference. By leveraging modern advances in deep probabilistic programming languages, researchers can easily scale these models up to larger, naturalistic datasets. Additionally, the lowered cost of model-building will accelerate rigorous theory-building and hypothesis-testing between competing theories of emotion.

The target audience comprises two groups of researchers. The first group includes researchers, such as cognitive psychologists, who favor theoretically-grounded models of affective phenomena, and who wish to scale their models up to complex, naturalistic datasets. The second group includes computer scientists, especially those who primarily use deep learning, who wish to add more emotion theory into their deep learning models, and in a principled manner. Deep probabilistic programming offers a way to combine the benefits of these two approaches to affective computing.

We will be learning from a webbook, using Jupyter notebooks. We will start with introductory primers to probabilistic programming concepts, such as stochastic primitives; compositionality and recursion; and stochastic variational inference. We will then transition into worked examples on previously-collected affective computing datasets. We will be using the open-sourced deep probabilistic programming language Pyro, first released in 2017. Tutorial participants will be able to download and run the code on their local machines as they follow along the material. We hope that by the end of this short tutorial, participants will be inspired—and equipped with some basic skills—to adopt approaches like deep probabilistic programming that merge theory- and data-driven approaches, and that such efforts will lead to greater collaboration between these historically-distinct paradigms in affective computing.

Edit: We will now be using Google Colaboratory (Colab), which provides a Jupyter environment on the cloud, hosted by Google. This means that participants do not need to download any code onto their computers to follow along.

Table of Contents

Affective Computing relevant examples:

  • Example 1 [Colab Notebook]: Linear Regression as a model of Appraisal.
  • Example 2 [Colab Notebook]: Variational Autoencoder to generate emotional faces.
  • Example 3 Colab Notebook: Semi-supervised VAE to learn to recognize emotions from faces. (Currently buggy, sorry!)
  • Example 4 [Colab Notebook]: Multimodal VAE to model latent affect.

These examples were tested on Pyro version 0.4.1



Acknowledgments

This work was supported indirectly by a number of grants:

  • the A*STAR Human-Centric Artificial Intelligence Programme (SERC SSF Project No. A1718g0048)
  • Singapore MOE AcRF Tier 1 (No. 251RES1709) to HS
  • NIH Grant 1R01MH112560-01 to JZ.
  • This material is based on research sponsored by DARPA under agreement number FA8750-14-2-0009.