Datasets:

Modalities:
Audio
Video
ArXiv:
Libraries:
Datasets
License:
Dataset Viewer (First 5GB)
Auto-converted to Parquet Duplicate
Search is not available for this dataset
The dataset viewer is not available for this split.
Rows from parquet row groups are too big to be read: 1.19 GiB (max=286.10 MiB)
Error code:   TooBigContentError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

ResponseNet

ResponseNet is a large-scale dyadic video dataset designed for Online Multimodal Conversational Response Generation (OMCRG). It fills the gap left by existing datasets by providing high-resolution, split-screen recordings of both speaker and listener, separate audio channels, and word‑level textual annotations for both participants.

Paper

If you use this dataset, please cite:

ResponseNet: A High‑Resolution Dyadic Video Dataset for Online Multimodal Conversational Response Generation
Authors: Luo, Cheng and Wang, Jianghui and Li, Bing and Song, Siyang and Ghanem, Bernard

Github Project

Features

  • 696 temporally synchronized dyadic video pairs (over 14 hours total).
  • High-resolution (1024×1024) frontal‑face streams for both speaker and listener.
  • Separate audio channels for fine‑grained verbal and nonverbal analysis.
  • Word‑level textual annotations for both participants.
  • Longer clips (average 73.39 s) than REACT2024 (30 s) and Vico (9 s), capturing richer conversational exchanges.
  • Diverse topics: professional discussions, emotionally driven interactions, educational settings, interdisciplinary expert talks.
  • Balanced splits: training, validation, and test sets with equal distributions of topics, speaker identities, and recording conditions.

Data Fields

Each example in the dataset is a dictionary with the following fields:

  • video/speaker: Path to the speaker’s video stream (1024×1024, frontal view).
  • video/listener: Path to the listener’s video stream (1024×1024, frontal view).
  • audio_speaker: Path to the speaker’s separated audio channel.
  • audio/listener: Path to the listener’s separated audio channel.
  • transcript/speaker: Word‑level transcription for the speaker (timestamps included).
  • transcript/listener: Word‑level transcription for the listener (timestamps included).
  • vector/speaker: Path to the speaker’s facial attributes.
  • vector/listener: Path to the listener’s facial attributes.

Dataset Splits

We follow a standard 6:2:2 split ratio, ensuring balanced distributions of topics, identities, and recording conditions:

Split # Video Pairs Proportion (%)
Train 417 59.9
Valid 139 20.0
Test 140 20.1
Total 696 100.0

Visualization

You can visualize word‑cloud statistics, clip‑duration distributions, and topic breakdowns using standard Python plotting tools.

Citation

@article{luo2025omniresponse,
  title={OmniResponse: Online Multimodal Conversational Response Generation in Dyadic Interactions},
  author={Luo, Cheng and Wang, Jianghui and Li, Bing and Song, Siyang and Ghanem, Bernard},
  journal={arXiv preprint arXiv:2505.21724},
  year={2025}
}}

License

This dataset is released under the CC BY-NC 4.0 license.

Downloads last month
207