(Licensed under the MIT License)
For details on the associated research, please refer to the paper:
Hassan Hayat, Carles Ventura, Agata Lapedriza, “Recognizing Emotions Evoked by Movies Using Multitask Learning”
International Conference on Affective Computing & Intelligent Interaction (ACII 2021)
This package was developed by Mr. Hassan Hayat (hhassan0@uoc.edu). For any inquiries regarding this package, please feel free to contact him. The package is provided free for academic use and is distributed at your own risk.
- Ubuntu Linux
- Python: 3.x.x
- GPU: With CUDA support
- Visual Features: I3D Model
- TensorFlow: 1.14
- COGNIMUSE Dataset: Access Here
The COGNIMUSE dataset provides subtitle information along with timestamps, which indicate the starting and ending frames during which each subtitle appears. These timestamps are used to create corresponding movie clips.
An I3D model is utilized to extract visual features from the movie clips. Specifically, the output from the 'Mixed-5c' layer serves as the visual representation of each frame. Prior to input into the model, each frame is resized to 224x224 pixels.
-
Text Modality:
To train, validate, and test using only the text modality, execute:./text_modality/st_main.py
-
Visual Modality:
To train, validate, and test using only the visual modality, execute:./visual_modality/st_main.py
- Text + Visual Modalities:
To train, validate, and test using both text and visual modalities, execute:./multi_modality/st_main.py
-
Text Modality:
To train, validate, and test using only the text modality in a multi-task setting, execute:./text_modality/mt_main.py
-
Visual Modality:
To train, validate, and test using only the visual modality in a multi-task setting, execute:./visual_modality/mt_main.py
- Text + Visual Modalities:
To train, validate, and test using both text and visual modalities in a multi-task setting, execute:./multi_modality/mt_main.py