--- license: cc-by-nc-4.0 --- # ResponseNet **ResponseNet** is a large-scale dyadic video dataset designed for **Online Multimodal Conversational Response Generation (OMCRG)**. It fills the gap left by existing datasets by providing high-resolution, split-screen recordings of both speaker and listener, separate audio channels, and word‑level textual annotations for both participants. ## Paper If you use this dataset, please cite: > **ResponseNet: A High‑Resolution Dyadic Video Dataset for Online Multimodal Conversational Response Generation (Neurips 2025)** > *Authors: Luo, Cheng and Wang, Jianghui and Li, Bing and Song, Siyang and Ghanem, Bernard* [Github](https://github.com/awakening-ai/OmniResponse) [Project](https://omniresponse.github.io/) ## Features - **696** temporally synchronized dyadic video pairs (over **14 hours** total). - **High-resolution** (1024×1024) frontal‑face streams for both speaker and listener. - **Separate audio channels** for fine‑grained verbal and nonverbal analysis. - **Word‑level textual annotations** for both participants. - **Longer clips** (average **73.39 s**) than REACT2024 (30 s) and Vico (9 s), capturing richer conversational exchanges. - **Diverse topics**: professional discussions, emotionally driven interactions, educational settings, interdisciplinary expert talks. - **Balanced splits**: training, validation, and test sets with equal distributions of topics, speaker identities, and recording conditions. ## Data Fields Each example in the dataset is a dictionary with the following fields: - `video/speaker`: Path to the speaker’s video stream (1024×1024, frontal view). - `video/listener`: Path to the listener’s video stream (1024×1024, frontal view). - `audio_speaker`: Path to the speaker’s separated audio channel. - `audio/listener`: Path to the listener’s separated audio channel. - `transcript/speaker`: Word‑level transcription for the speaker (timestamps included). - `transcript/listener`: Word‑level transcription for the listener (timestamps included). - `vector/speaker`: Path to the speaker’s facial attributes. - `vector/listener`: Path to the listener’s facial attributes. ## Dataset Splits We follow a standard **6:2:2** split ratio, ensuring balanced distributions of topics, identities, and recording conditions: | Split | # Video Pairs | Proportion (%) | |------------|---------------|----------------| | **Train** | 417 | 59.9 | | **Valid** | 139 | 20.0 | | **Test** | 140 | 20.1 | | **Total** | 696 | 100.0 | ## Visualization You can visualize word‑cloud statistics, clip‑duration distributions, and topic breakdowns using standard Python plotting tools. ## Citation ```bibtex @article{luo2025omniresponse, title={OmniResponse: Online Multimodal Conversational Response Generation in Dyadic Interactions}, author={Luo, Cheng and Wang, Jianghui and Li, Bing and Song, Siyang and Ghanem, Bernard}, journal={arXiv preprint arXiv:2505.21724}, year={2025} }} ``` ## License This dataset is released under the **CC BY-NC 4.0** license.