iseg.code icon indicating copy to clipboard operation
iseg.code copied to clipboard

This repo is the official implementation of iSeg: An Iterative Refinement-based Framework for Training-free Segmentation. Now We provide an interactive demo and the other part will be released as soon...

iSeg: An Iterative Refinement-based Framework for Training-free Segmentation

This repo is the official implementation of iSeg: An Iterative Refinement-based Framework for Training-free Segmentation

Introduction

  • We present a simple iterative training-free segmentation framework using stable diffusion, named iSeg. Based on the self attention maps in stable diffusion model, iSeg provides a more and more accurate semantic segmentation result with the increasing iterations.
  • We introduce an entropy-reduced self-attention module to remove the diffusion of the irrelevant global information in original self attention maps.
  • We further introduce a category-enhanced cross-attention module for the more accurate initial cross attention.
  • iSeg is powerful enough for zero-shot and open-vocabulary segmentation tasks, which can be used for not only natural and synthetic images, but also clipart, painting, sketch images and so on.
  • We further provide interaction demo which is able to segment objects by points, lines, boxes and descriptions.

For further details, please check out our paper.

Installation

following the codeblock below to install the dependencies.

# create conda environment
conda env create -f environment.yaml
conda activate iSeg

We use stable diffusion model 2.1-base as our base model, which is available in here.

Our code will automatically download the model. If you failed to download the model because of proxy, you can try to add the following code at the beginning of interactive_iSeg.py.

import os
os.environ['HF_ENDPOINT'] = 'https://hf-mirror.com'

Usage

this work is time-efficient and memory-efficient and without training, which means you can run it on a single GPU with 5GB memory and directly run it as follows

Interactive Demo

We provide an interactive demo for you to try our model. The demo can provide segmentation results by points, lines, boxes and descriptions. To run it, please follow the steps below:

cd ui
sh run.sh
interaction

Weakly Supervised Semantic Segmentation

TODO

Open-Vocabulary Semantic Segmentation

TODO

Results

TODO

Citation

@article{Lin_2024_iSeg,
    author    = {Lin Sun and Jiale Cao and Jin Xie and Fahad Shahbaz Khan and Yanwei Pang   },
    title     = {iSeg: An Iterative Refinement-based Framework for Training-free Segmentation},
    journal   = {arXiv preprint arXiv:2409.03209},
    year      = {2024},
}

Acknowledgement

We would like to acknowledge the contributions of public projects, such as SLiMe, TagCLIP whose code has been utilized in this repository.