This is the official implementation of the paper Take an Emotion Walk: Perceiving Emotions from Gaits Using Hierarchical Attention Pooling and Affective Mapping. Please add the following citation in your work if you use our code:
@InProceedings{taew, author = {Bhattacharya, Uttaran and Roncal, Christian and Mittal, Trisha and Chandra, Rohan and Bera, Aniket and Manocha, Dinesh}, title = {Take an Emotion Walk: Perceiving Emotions from Gaits Using Hierarchical Attention Pooling and Affective Mapping}, booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)}, month = {August}, year = {2020} }
Our scripts have been tested on Ubuntu 18.04 LTS with
- Python 3.6
- Cuda 10.2
- cudNN 7.6.5
We recommend using an Anaconda virtual environment. If Anaconda is not already installed, Install Anaconda and run
conda env create -n taew -f environment.yml
from within the project directory
Run the following command from within the project directory to download and extract the sample datasets and network weights:
sh download_data_weights.sh
We have used the Emotion-Gait dataset for this work. The full dataset is available for download here: https://go.umd.edu/emotion-gait.
- Activate the conda environment
conda activate taew
- Run the evaluation script For dgnn evaluation:
python evaluate.py --dgnn
For stgcn evaluation:
python evaluate.py --stgcn
For lstm network evaluation:
python evaluate.py --lstm
For step evaluation:
python evaluate.py --step
For taew evaluation:
python evaluate.py --taew
main.py
is the starting point of the code. It is runnable out-of-the-box once the datasets directory is downloaded and extracted. It also contains the full list of arguments for using the code.utils/loader.py
is used for loading the data and the labels. Labels are only available for the annotated part of the data.utils/processor.py
contains the main training routine with forward and backward passes on the network, and parameter updates per iteration.net/hapy.py
contains the overall network and description of the forward pass.