The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError
Exception: FileNotFoundError
Message: tencent/VCB-Bench@39c3387c65423c4218329f3d68846d4ce41ce416/vcb_bench/pretrain_evaluation/story_continue/audio/context_sc_common_000001.wav (repository not found)
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
response.raise_for_status()
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/api/datasets/tencent/VCB-Bench/revision/39c3387c65423c4218329f3d68846d4ce41ce416
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 125, in _repo_and_revision_exist
self._api.repo_info(
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2816, in repo_info
return method(
^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2674, in dataset_info
hf_raise_for_status(r)
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 459, in hf_raise_for_status
raise _format(RepositoryNotFoundError, message, response) from e
huggingface_hub.errors.RepositoryNotFoundError: 401 Client Error. (Request ID: Root=1-694bd4b2-6b1889ea2fb28d284edf027a;0dccb3a0-4f1b-4a48-85d5-2a081774f1dd)
Repository Not Found for url: https://huggingface.co/api/datasets/tencent/VCB-Bench/revision/39c3387c65423c4218329f3d68846d4ce41ce416.
Please make sure you specified the correct `repo_id` and `repo_type`.
If you are trying to access a private or gated repo, make sure you are authenticated. For more details, see https://huggingface.co/docs/huggingface_hub/authentication
Invalid username or password.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1595, in _prepare_split_single
num_examples, num_bytes = writer.finalize()
^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 728, in finalize
self.write_examples_on_file()
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 581, in write_examples_on_file
self.write_batch(batch_examples=batch_examples)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 701, in write_batch
self.write_table(pa_table, writer_batch_size)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 716, in write_table
pa_table = embed_table_storage(pa_table)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2249, in embed_table_storage
embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1795, in wrapper
return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2124, in embed_array_storage
return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 291, in embed_storage
(path_to_bytes(x["path"]) if x["bytes"] is None else x["bytes"]) if x is not None else None
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/py_utils.py", line 310, in wrapper
return func(value) if value is not None else None
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 286, in path_to_bytes
with xopen(path, "rb", download_config=download_config) as f:
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 935, in xopen
file_obj = fsspec.open(file, mode=mode, *args, **kwargs).open()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 135, in open
return self.__enter__()
^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 103, in __enter__
f = self.fs.open(self.path, mode=mode)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "<string>", line 3, in open
File "/usr/local/lib/python3.12/unittest/mock.py", line 1139, in __call__
return self._mock_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/unittest/mock.py", line 1143, in _mock_call
return self._execute_mock_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/unittest/mock.py", line 1204, in _execute_mock_call
result = effect(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 770, in wrapped
f = fs_open(self, urlpath, mode, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1293, in open
f = self._open(
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 275, in _open
return HfFileSystemFile(self, path, mode=mode, revision=revision, block_size=block_size, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 938, in __init__
self.resolved_path = fs.resolve_path(path, revision=revision)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 200, in resolve_path
_raise_file_not_found(path, err)
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1138, in _raise_file_not_found
raise FileNotFoundError(msg) from err
FileNotFoundError: tencent/VCB-Bench@39c3387c65423c4218329f3d68846d4ce41ce416/vcb_bench/pretrain_evaluation/story_continue/audio/context_sc_common_000001.wav (repository not found)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1334, in compute_config_parquet_and_info_response
parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 911, in stream_convert_to_parquet
builder._prepare_split(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1447, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1604, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
audio
audio |
|---|
Introduction
Voice Chat Bot Bench (VCB Bench) is a high-quality Chinese benchmark built entirely on real human speech. It evaluates large audio language models (LALMs) along three complementary dimensions:
(1) Instruction following: Text Instruction Following (TIF), Speech Instruction Following (SIF), English Text Instruction Following (TIF-En), English Speech Instruction Following (SIF-En) and Multi-turn Dialog (MTD);
(2) Knowledge: General Knowledge (GK), Mathematical Logic (ML), Discourse Comprehension (DC) and Story Continuation (SC).
(3) Robustness: Speaker Variations (SV), Environmental Variations (EV), and Content Variations (CV).
Getting Started
Installation:
git clone https://github.com/Tencent/VCB-Bench.git
cd VCB-Bench
pip install -r requirements.txt
Note: To evaluate Qwen3-omni, please replace it with the environment it requires.
Download Dataset:
Download the dataset from Hugging Face and place the 'vcb_bench' into 'data/downloaded_datasets'.
Evaluation:
This code is adapted from Kimi-Audio-Evalkit, where you can find more details about the evaluation commands.
(1) Inference + Evaluation:
python run_audio.py --model {model_name} --data {data_name}
For example:
CUDA_VISIBLE_DEVICES=1 python run_audio.py --model Qwen2.5-Omni-7B --data general_knowledge
(2) Only Inference:
python run_audio.py --model {model_name} --data {data_name} --skip-eval
For example:
CUDA_VISIBLE_DEVICES=4,5,6,7 python run_audio.py --model StepAudio --data continuation_en creation_en empathy_en recommendation_en rewriting_en safety_en simulation_en emotional_control_en language_control_en non_verbal_vocalization_en pacing_control_en style_control_en volume_control_en --skip-eval
(3) Only Evaluation:
python run_audio.py --model {model_name} --data {data_name} --reeval
For example:
CUDA_VISIBLE_DEVICES=2 nohup python run_audio.py --model Mimo-Audio --data continuation creation empathy --reeval
(4) Inference + ASR + Evaluation:
python run_audio.py --model {model_name} --data {data_name} --wasr
For example:
CUDA_VISIBLE_DEVICES=3 python run_audio.py --model StepAudio2 --data rewriting safety simulation continuation_en --wasr
Format Result:
python sumup_eval.py --model {model_name}
python sumup_eval.py --model {model_name} --export_excel --output_file my_results.xlsx
Supported Datasets and Models
(1) Locate the dataset you need to evaluate from the Data Name column in the Datasets table, and populate the {data_name} parameter in the evaluation command accordingly.
(2) Each dataset in the SV, EV, and CV sections has a corresponding comparison dataset named "{data_name}_cmp", following the specified naming convention.
(3) Identify the model you intend to evaluate from the Model Name column in the Models table, and insert the appropriate {model_name} into the evaluation command.
Datasets:
| Data Type | Data Name | Detail |
|---|---|---|
| TIF | continuation | - |
| creation | - | |
| empathy | - | |
| recommendation | - | |
| rewriting | - | |
| safety | - | |
| simulation | - | |
| TIF-En | continuation_en | - |
| creation_en | - | |
| empathy_en | - | |
| recommendation_en | - | |
| rewriting_en | - | |
| safety_en | - | |
| simulation_en | - | |
| SIF | emotional_control | - |
| language_control | - | |
| non_verbal_vocalization | - | |
| pacing_control | - | |
| style_control | - | |
| volume_control | - | |
| SIF-En | emotional_control_en | - |
| language_control_en | - | |
| non_verbal_vocalization_en | - | |
| pacing_control_en | - | |
| style_control_en | - | |
| volume_control_en | - | |
| MTD | progression | - |
| backtracking | - | |
| transition | - | |
| GK | general_knowledge | mathematics, geography, politics, chemistry, biology, law, physics, history, medicine, economics, sports, culture |
| ML | basic_math | - |
| math | - | |
| logical_reasoning | analysis, induction, analogy, logic | |
| DC | discourse_comprehension | inference, induction, analysis |
| SV | age | child, elder |
| accent | tianjin, beijing, dongbei, sichuan | |
| volume | down, up | |
| speed | - | |
| EV | non_vocal_noise | echo, outdoors, far_field |
| vocal_noise | TV_playback, background_chat, vocal_music, voice_announcement | |
| unstable_signal | - | |
| CV | casual_talk | - |
| mispronunciation | - | |
| grammatical_error | - | |
| topic_shift | - | |
| code_switching | - |
Models:
| Model Type | Model Name |
|---|---|
| Chat Model | Qwen2-Audio-7B-Instruct |
| Qwen2.5-Omni-7B | |
| Baichuan-Audio-Chat | |
| GLM4-Voice | |
| Kimi-Audio | |
| Mimo-Audio | |
| StepAudio | |
| StepAudio2 | |
| GPT4O-Audio | |
| Qwen3-Omni-Instruct | |
| Pretrain Model | Qwen2-Audio-7B |
| Baichuan-Audio | |
| Kimi-Audio-Base | |
| StepAudio2-Base |
Acknowledge
We borrow some code from Kimi-Audio-Evalkit, GLM-4-Voice, Baichuan-Audio, Kimi-Audio, Mimo-Audio, Step-Audio2, and StepAudio.
Citation
@misc{hu2025vcbbenchevaluationbenchmark,
title={VCB Bench: An Evaluation Benchmark for Audio-Grounded Large Language Model Conversational Agents},
author={Jiliang Hu and Wenfu Wang and Zuchao Li and Chenxing Li and Yiyang Zhao and Hanzhao Li and Liqiang Zhang and Meng Yu and Dong Yu},
year={2025},
eprint={2510.11098},
archivePrefix={arXiv},
primaryClass={cs.SD},
url={https://arxiv.org/abs/2510.11098},
}
- Downloads last month
- -