Code and results data underlying Chapter 4 of the PhD thesis: "Towards predicting memory in multimodal group interactions"
DOI: 10.4121/1ac2b163-f9f5-4df0-9485-dc80ee3b632f
Datacite citation style
Dataset
This repository contains the complete code and experiment outputs for Chapter 4 of M. Tsfasman's PhD thesis "Towards predicting memory in multimodal group interactions". It provides Python scripts and notebooks for preprocessing multimodal data, training and evaluating machine learning models to predict memorable moments in conversations using features such as eye-gaze and speaker activity, and performing feature ablation studies to assess the importance of each input. The repository includes robust session-based cross-validation, hyper-parameter optimization, and tools for result visualization. It is designed for reproducible research and is an extended, methodologically improved version of a published ICMI 2022 paper*, supporting both local and cluster-based execution.
*M Tsfasman, K Fenech, M Tarvirdians, A Lorincz, C Jonker, C Oertel, "Towards creating a conversational memory for long-term meeting support: predicting memorable moments in multi-party conversations through eye-gaze,'' in Proc. International Conference on Multimodal Interaction (ICMI)}, pp. 94–104, 2022.
History
- 2025-07-25 first online, published, posted
Publisher
4TU.ResearchDataFormat
.py, .sh, .csv, .ipynb, .err, .out, .txt, .yml, .sif, .tex, .mdOrganizations
TU Delft, Faculty of Electrical Engineering, Mathematics and Computer Science, Department of Intelligent Systems, Interactive IntelligenceDATA
Files (1)
- 6,872,101,464 bytesMD5:
81e5951279b8f20a62038b830c93a9a2
Chapter4_code(ICMI_extension).zip