Skip to content

The PyTorch Code for our paper "Dual Progressive Transformations for Weakly Supervised Semantic Segmentation"

Notifications You must be signed in to change notification settings

huodongjian0603/crt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CRT

Official PyTorch implementation for our paper "Dual Progressive Transformations for Weakly Supervised Semantic Segmentation" [paper]

Abstract

Weakly supervised semantic segmentation (WSSS), which aims to mine the object regions by merely using class-level labels, is a challenging task in computer vision. The current state-of-the-art CNN-based methods usually adopt Class-Activation-Maps (CAMs) to highlight the potential areas of the object, however, they may suffer from the part-activated issues. To this end, we try an early attempt to explore the global feature attention mechanism of vision transformer in WSSS task. However, since the transformer lacks the inductive bias as in CNN models, it can not boost the performance directly and may yield the over-activated problems. To tackle these drawbacks, we propose a Convolutional Neural Networks Refined Transformer (CRT) to mine a globally complete and locally accurate class activation maps in this paper. To validate the effectiveness of our proposed method, extensive experiments are conducted on PASCAL VOC 2012 and CUB-200-2011 datasets. Experimental evaluations show that our proposed CRT achieves the new state-of-the-art performance on both the weakly supervised semantic segmentation task the weakly supervised object localization task, which outperform others by a large margin.

Requirement

  • Python 3.7
  • PyTorch 1.1.0+
  • NVIDIA GeForce RTX 2080Ti x 2

Usage

Preparation

  1. Download the repository.
git clone https://github.com/huodongjian0603/crt.git
  1. Install dependencies.
pip install -r requirements.txt
  1. Download PASCAL VOC 2012 devkit.

Generate pseudo-segmentation labels

  1. Run script run_sample.py.
python run_sample.py --voc12_root $downloaded_dataset_path/VOCdevkit/VOC2012

After the script completes, pseudo labels are generated in the following directory and their quality is evaluated in mIoU. If you want to train DeepLab, add --infer_list voc12/train_aug.txt to the above script. The former and the latter respectively generate 1464 and 10582 pseudo-segmentation masks in .png format in the .\result\seg_sem.

.
├── misc
├── net
├── result  # generated cam and pseudo labels
│   ├── cam
│   ├── ins_seg
│   ├── ir_label
│   └── sem_seg # what we want in this step!
├── sess  # saved models
│   ├── deits_cam.pth
│   ├── res152_cam.pth
│   └── res50_irn.pth
├── step
├── voc12
├── requirements.txt
├── run_sample.py
└── sample_train_eval.log
  1. Move .\result\sem_seg to $downloaded_dataset_path/VOCdevkit/VOC2012, and rename it to pseudo_seg_labels. You can actually move the folder, or make a soft link(recommanded).
ln -s .\result\sem_seg $downloaded_dataset_path/VOCdevkit/VOC2012\pseudo_seg_labels

The file structure of VOC2012 should look like this:

VOC2012
├─Annotations
├─ImageSets
│  ├─Action
│  ├─Layout
│  ├─Main
│  └─Segmentation
├─JPEGImages
├─SegmentationClass
├─SegmentationObject
└─pseudo_seg_labels

Train DeepLab with the generated pseudo labels.

  1. Change the working directory to deeplab/. Download the pretrained models and put them into the pretrained folder.
cd deeplab
  1. Modify the configuration file ./configs/voc12_resnet_dplv2.yaml.
DATASET:
    NAME: vocaug
    ROOT: ./../../VOC2012  # Change the directory to where your VOC2012 is located
    LABELS: ./data/datasets/voc12/labels.txt
    N_CLASSES: 21
    IGNORE_LABEL: 255
    SCALES: [0.5, 0.75, 1.0, 1.25, 1.5]
    SPLIT:
        TRAIN: train_aug
        VAL: val
        TEST: test
  1. Train DeepLabv2-resnet101 model.
python main.py train \
      --config-path configs/voc12_resnet_dplv2.yaml
  1. Evaluate DeepLabv2-resnet101 model on the validation set.
python main.py test \
    --config-path configs/voc12_resnet_dplv2.yaml \
    --model-path data/models/voc12/voc12_resnet_v2/train_aug/checkpoint_final.pth
  1. Re-evaluate with a CRF post-processing.
python main.py crf \
    --config-path configs/voc12_resnet_dplv2.yaml

(optional) Exploratory experiments on weakly supervised object localization (WSOL) tasks.

We found that the proposed CRT method is equally suitable for the WSOL task, and only a simple ResNet50 modification of the Deit-S branch can achieve promising results (without the improvement for Deit-S in the WSSS task). Here, we provide a naive implementation for WSOL task. You just need to follow the instructions of TS-CAM and replace some files in TS-CAM with the files we provide (see step 2 below) to achieve the results in the paper.

For those who are lazy(LOL), we also provide a simple tutorial here, but we still strongly recommend browsing the TS-CAM repository for details.

  1. Download the repository.
git clone https://github.com/vasgaowei/TS-CAM.git
  1. Replace the folder with the same name in TS-CAM/ with the folder in wsol/backup/
wsol
├─backup
│  ├─configs
│  ├─lib  # main diff with TS-CAM: ResNet50_cam
│  └─tools_cam  # main diff with TS-CAM: train_cam.py
├─ckpt
└─log
  1. Configure the dataset path in file deit_tscam_small_patch16_224.yaml
DATA:
  DATASET: CUB
  DATADIR: data/CUB_200_2011 # change your path here
  NUM_CLASSES: 200
  RESIZE_SIZE : 256
  CROP_SIZE : 224
  IMAGE_MEAN : [0.485, 0.456, 0.406]
  IMAGE_STD : [0.229, 0.224, 0.225]
  1. Training.
bash train_val_cub.sh 0,1 deit small 224
  1. Evaluation.
bash val_cub.sh 0 deit small 224 ${MODEL_PATH}

Performance

Quality

Visualization of pseudo-segmentation masks on the PASCAL VOC 2012 training set.
a) Input image; b) Ground truth; c) IRNet; d) TS-CAM; e) CRT


Visualization of pseudo-segmentation masks on the PASCAL VOC 2012 val set.
a) Input image; b) Ground truth; c) CRT

Quantity

Pseudo segmentation mask

Dataset Seed Mask Weight
PASCAL VOC 57.7 71.8 Download

WSSS results

Dataset Val Test Weight
PASCAL VOC 71.2 71.7 Download

WSOL results

Dataset Top-1 Top-5 Gt-Known Weight
CUB-200-2011 72.9 86.4 90.1 Download

TODO

  • complete test

Citation

If you find the code useful, please consider citing our paper using the following BibTeX entry.

@misc{
  https://doi.org/10.48550/arxiv.2209.15211,
  doi = {10.48550/ARXIV.2209.15211},
  url = {https://arxiv.org/abs/2209.15211},
  author = {Huo, Dongjian and Su, Yukun and Wu, Qingyao},
  keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {Dual Progressive Transformations for Weakly Supervised Semantic Segmentation},
  publisher = {arXiv},
  year = {2022},
  copyright = {arXiv.org perpetual, non-exclusive license}
}

Acknowledgement

Our project references the codes in the following repos.

About

The PyTorch Code for our paper "Dual Progressive Transformations for Weakly Supervised Semantic Segmentation"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages