ReasonSeg_test / README.md
Ricky06662's picture
Improve dataset card: Add task category, license, paper, and code links for ReasonSeg-Test (#2)
5153694 verified
---
license: cc-by-nc-4.0
task_categories:
- image-segmentation
language:
- en
tags:
- reasoning
- zero-shot
- reinforcement-learning
- multi-modal
- VLM
size_categories:
- n<1K
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
- name: mask
sequence:
sequence: bool
- name: image_id
dtype: string
- name: ann_id
dtype: string
- name: img_height
dtype: int64
- name: img_width
dtype: int64
splits:
- name: test
num_bytes: 1666685613.0
num_examples: 779
download_size: 1235514015
dataset_size: 1666685613.0
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
---
# ReasonSeg Test Dataset
This repository contains the **ReasonSeg Test Dataset**, which serves as an evaluation benchmark for the paper [Seg-Zero: Reasoning-Chain Guided Segmentation via Cognitive Reinforcement](https://arxiv.org/abs/2503.06520).
**Code:** [https://github.com/dvlab-research/Seg-Zero](https://github.com/dvlab-research/Seg-Zero)
## Paper Abstract
Traditional methods for reasoning segmentation rely on supervised fine-tuning with categorical labels and simple descriptions, limiting its out-of-domain generalization and lacking explicit reasoning processes. To address these limitations, we propose Seg-Zero, a novel framework that demonstrates remarkable generalizability and derives explicit chain-of-thought reasoning through cognitive reinforcement. Seg-Zero introduces a decoupled architecture consisting of a reasoning model and a segmentation model. The reasoning model interprets user intentions, generates explicit reasoning chains, and produces positional prompts, which are subsequently used by the segmentation model to generate precious pixel-level masks. We design a sophisticated reward mechanism that integrates both format and accuracy rewards to effectively guide optimization directions. Trained exclusively via reinforcement learning with GRPO and without explicit reasoning data, Seg-Zero achieves robust zero-shot generalization and exhibits emergent test-time reasoning capabilities. Experiments show that Seg-Zero-7B achieves a zero-shot performance of 57.5 on the ReasonSeg benchmark, surpassing the prior LISA-7B by 18%. This significant improvement highlights Seg-Zero's ability to generalize across domains while presenting an explicit reasoning process.
## About Seg-Zero
Seg-Zero is a novel framework for reasoning segmentation that utilizes cognitive reinforcement to achieve remarkable generalizability and explicit chain-of-thought reasoning.
<div align=center>
<img width="98%" src="https://github.com/dvlab-research/Seg-Zero/raw/main/assets/overview.png"/>
</div>
Seg-Zero demonstrates the following features:
1. Seg-Zero exhibits emergent test-time reasoning ability. It generates a reasoning chain before producing the final segmentation mask.
2. Seg-Zero is trained exclusively using reinforcement learning, without any explicit supervised reasoning data.
3. Compared to supervised fine-tuning, our Seg-Zero achieves superior performance on both in-domain and out-of-domain data.
### Model Pipeline
Seg-Zero employs a decoupled architecture, including a reasoning model and segmentation model. A sophisticated reward mechanism integrates both format and accuracy rewards.
<div align=center>
<img width="98%" src="https://github.com/dvlab-research/Seg-Zero/raw/main/assets/pipeline.png"/>
</div>
### Examples
<div align=center>
<img width="98%" src="https://github.com/dvlab-research/Seg-Zero/raw/main/assets/examples.png"/>
</div>
## Sample Usage: Evaluation
This dataset (`ReasonSeg-Test`) is designed for evaluating the zero-shot performance of models like Seg-Zero on reasoning-based image segmentation tasks.
First, install the necessary dependencies for the Seg-Zero project:
```bash
git clone https://github.com/dvlab-research/Seg-Zero.git
cd Seg-Zero
conda create -n visionreasoner python=3.12
conda activate visionreasoner
pip install torch==2.6.0 torchvision==0.21.0
pip install -e .
```
Then, you can run evaluation using the provided scripts. Make sure to download pretrained models first:
```bash
mkdir pretrained_models
cd pretrained_models
git lfs install
git clone https://huggingface.co/Ricky06662/VisionReasoner-7B
```
With the pretrained models downloaded, you can run the evaluation script for ReasonSeg:
```bash
bash evaluation_scripts/eval_reasonseg_visionreasoner.sh
```
Adjust `'--batch_size'` in the bash scripts based on your GPU. You will see the gIoU in your command line.
<div align=center>
<img width="98%" src="https://github.com/dvlab-research/Seg-Zero/raw/main/assets/val_results.png"/>
</div>
## The GRPO Algorithm
Seg-Zero generates several samples, calculates the rewards and then optimizes towards samples that achieve higher rewards.
<div align=center>
<img width="48%" src="https://github.com/dvlab-research/Seg-Zero/raw/main/assets/rl_sample.png"/>
</div>
## Citation
If you use this dataset or the Seg-Zero framework, please cite the associated papers:
```bibtex
@article{liu2025segzero,
title = {Seg-Zero: Reasoning-Chain Guided Segmentation via Cognitive Reinforcement},
author = {Liu, Yuqi and Peng, Bohao and Zhong, Zhisheng and Yue, Zihao and Lu, Fanbin and Yu, Bei and Jia, Jiaya},
journal = {arXiv preprint arXiv:2503.06520},
year = {2025}
}
@article{liu2025visionreasoner,
title = {VisionReasoner: Unified Visual Perception and Reasoning via Reinforcement Learning},
author = {Liu, Yuqi and Qu, Tianyuan and Zhong, Zhisheng and Peng, Bohao and Liu, Shu and Yu, Bei and Jia, Jiaya},
journal = {arXiv preprint arXiv:2505.12081},
year = {2025}
}
```