demo_reflex.mp4
REFLEX (Robotic Explanations to FaiLures and Human EXpressions) is a comprehensive multimodal dataset capturing human reactions to robot failures and subsequent explanations in collaborative settings. This dataset facilitates research into human-robot interaction dynamics, addressing the need to study reactions to both initial failures and explanations, as well as the evolution of these reactions in long-term interactions.
The data was collected from a user study where participants collaborated with a robot on a task where the robot experienced programmed failures. Different explanation strategies were employed across participants to study their effects.
├── Dataset
│ ├── Data # Download the dataset here
│ └── DatasetGuide.md # Info and Guides for the dataset
├── reflex_visualization/ # Code for visualizing the multimodal data
│ ├── requirements.txt # Python dependencies for visualization
│ └── src # Visualization Code
│ ├── ...
│ └── main.py # Main visualization script
├── LICENSE.md # MIT License details
└── README.md # This file
The repository includes a visualization tool built using Rerun, an open-source visualization sdk for multimodal data.
-
Clone this repository:
git clone https://github.com/andreasnaoum/reflex-viz.git
-
Download the Data from Zenodo and place them under the Dataset folder.
-
Install the dependencies for visualization:
cd reflex-visualization pip install -r requirements.txt
-
To visualize a participant interaction with the dataset:
python src/main.py --participant C1-1
Replace C1-1
with the participant code following the format {strategy}-{participant_number}
, where:
- Strategy is one of: C1, C2, C3, D1, or D2
- Participant number is between 1 and 11
See DatasetGuide.md
for more detailed about the dataset.
--participant
: Participant code in the format{strategy}-{number}
(e.g., 'C1-1')--max-frames
: Maximum number of frames to process (optional, default:None
)--jpeg-quality
: JPEG compression quality for images from 1-100 (optional, default: 15)--data-path
: Path to the data directory (optional)--face-3d
: Enable 3D face visualization (optional, default: false)--gaze-3d
: Enable 3D gaze visualization (optional, default: false)--body-3d
: Enable 3D body visualization (optional, default: false)--openface-confidence
: Minimum confidence threshold for OpenFace from 0.0-1.0 (optional, default: 0.7)
The visualization integrates multiple data modalities synchronized by time (or frame):
- Video from Camera 1 with overlaid landmarks (face, body, eyes)
- Current failure phase, conversation, gaze, and pose classification as text
- Graphs displaying emotion values, FAU intensities, and arousal scores
- Timeline of the interaction phases
The dataset allows for various analyses, including:
- Comparing human reactions to different robotic failures
- Assessing the effectiveness of different explanation strategies
- Studying how human satisfaction evolve with repeated failures
- Analyzing multimodal behavioral patterns in response to robot explanations
One interesting insight from initial analysis shows lower confusion likelihood during pick failures compared to more complex carry-place failures, suggesting pick failures were perceived as easier to resolve.
If you have any questions or issues using the dataset or visualization tools, please contact Andreas Naoum at anaoum@kth.se.
If you use our dataset, please cite our paper:
@inproceedings{reflex_dataset,
title={REFLEX Dataset: A Multimodal Dataset of Human Reactions to Robot Failures and Explanations},
author={Parag Khanna and Andreas Naoum and Elmira Yadollahi and Mårten Björkman and Christian Smith},
year={2025},
eprint={2502.14185},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2502.14185},
}