Skip to content

ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation

Notifications You must be signed in to change notification settings

UARK-AICV/ShapeFormer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation

Table of Contents

Introduction

alt text

The figure above illustrates our ShapeFormer architecture. The main implementation of this network can be found here.

Usage

1. Installation

This work uses aistron for implementation. Please follow the instruction at here

2. Data preparation

Please download the below datasets. More preparation instruction can be found here.

2.1. KINS dataset

Download the Images from KITTI dataset.

The Amodal Annotations could be found at KINS dataset

2.2. D2SA dataset

The D2S Amodal dataset could be found at mvtec-d2sa.

2.3. COCOA-cls dataset

The COCOA dataset annotation from here (reference from github.com/YihongSun/Bayesian-Amodal) The images of COCOA dataset is the train2014 and val2014 of COCO dataset.

3. Training, Testing and Demo

Configuration files for training AISFormer on each datasets are available here. To train, test and run demo, see the example scripts at scripts/:

Citation

@article{tran2024shapeformer,
  title={ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation},
  author={Tran, Minh and Bounsavy, Winston and Vo, Khoa and Nguyen, Anh and Nguyen, Tri and Le, Ngan},
  journal={arXiv preprint arXiv:2403.11376},
  year={2024}
}

About

ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published