python手势检测 cascade_Cascade R-CNN - 深入研究高质量的对象检测

韩夕
2023-12-01

Cascade R-CNN: Delving into High Quality Object Detection

by Zhaowei Cai and Nuno Vasconcelos

This repository is written by Zhaowei Cai at UC San Diego.

Introduction

This repository implements mulitple popular object detection algorithms, including Faster R-CNN, R-FCN, FPN, and our recently proposed Cascade R-CNN, on the MS-COCO and PASCAL VOC datasets. Multiple choices are available for backbone network, including AlexNet, VGG-Net and ResNet. It is written in C++ and powered by Caffe deep learning toolbox.

Cascade R-CNN is a multi-stage extension of the popular two-stage R-CNN object detection framework. The goal is to obtain high quality object detection, which can effectively reject close false positives. It consists of a sequence of detectors trained end-to-end with increasing IoU thresholds, to be sequentially more selective against close false positives. The output of a previous stage detector is forwarded to a later stage detector, and the detection results will be improved stage by stage. This idea can be applied to any detector based on the two-stage R-CNN framework, including Faster R-CNN, R-FCN, FPN, Mask R-CNN, etc, and reliable gains are available independently of baseline strength. A vanilla Cascade R-CNN on FPN detector of ResNet-101 backbone network, without any training or inference bells and whistles, achieved state-of-the-art results on the challenging MS-COCO dataset.

Update

The re-implementation of Cascade R-CNN in Detectron has been released. See Detectron-Cascade-RCNN. Very consistent improvements are available for all tested models, independent of baseline strength.

It is also recommended to use the third-party implementation, mmdetection based on PyTorch and tensorpack based on TensorFlow.

Citation

If you use our code/model/data, please cite our paper:

@inproceedings{cai18cascadercnn,

author = {Zhaowei Cai and Nuno Vasconcelos},

Title = {Cascade R-CNN: Delving into High Quality Object Detection},

booktitle = {CVPR},

Year = {2018}

}

or its extension:

@article{cai2019cascadercnn,

author = {Zhaowei Cai and Nuno Vasconcelos},

title = {Cascade R-CNN: High Quality Object Detection and Instance Segmentation},

journal = {arXiv preprint arXiv:1906.09756},

year = {2019}

}

Benchmarking

We benchmark mulitple detector models on the MS-COCO and PASCAL VOC datasets in the below tables.

MS-COCO (Train/Test: train2017/val2017, shorter size: 800 for FPN and 600 for the others)

model

#GPUs

bs

lr

iter

train time

test time

AP

AP50

AP75

VGG-RPN-baseline

2

4

3e-3

100k

12.5 hr

0.075s

23.6

43.9

23.0

VGG-RPN-Cascade

2

4

3e-3

100k

15.5 hr

0.115s

27.0

44.2

27.7

Res50-RFCN-baseline

4

1

3e-3

280k

19 hr

0.07s

27.0

48.7

26.9

Res50-RFCN-Cascade

4

1

3e-3

280k

22.5 hr

0.075s

31.1

49.8

32.8

Res101-RFCN-baseline

4

1

3e-3

280k

29 hr

0.075s

30.3

52.2

30.8

Res101-RFCN-Cascade

4

1

3e-3

280k

30.5 hr

0.085s

33.3

52.0

35.2

Res50-FPN-baseline

8

1

5e-3

280k

32 hr

0.095s

36.5

58.6

39.2

Res50-FPN-Cascade

8

1

5e-3

280k

36 hr

0.115s

40.3

59.4

43.7

Res101-FPN-baseline

8

1

5e-3

280k

37 hr

0.115s

38.5

60.6

41.7

Res101-FPN-Cascade

8

1

5e-3

280k

46 hr

0.14s

42.7

61.6

46.6

PASCAL VOC 2007 (Train/Test: 2007+2012trainval/2007test, shorter size: 600)

model

#GPUs

bs

lr

iter

train time

AP

AP50

AP75

Alex-RPN-baseline

2

4

1e-3

45k

2.5 hr

29.4

63.2

23.7

Alex-RPN-Cascade

2

4

1e-3

45k

3 hr

38.9

66.5

40.5

VGG-RPN-baseline

2

4

1e-3

45k

6 hr

42.9

76.4

44.1

VGG-RPN-Cascade

2

4

1e-3

45k

7.5 hr

51.2

79.1

56.3

Res50-RFCN-baseline

2

2

2e-3

90k

8 hr

44.8

77.5

46.8

Res50-RFCN-Cascade

2

2

2e-3

90k

9 hr

51.8

78.5

57.1

Res101-RFCN-baseline

2

2

2e-3

90k

10.5 hr

49.4

79.8

53.2

Res101-RFCN-Cascade

2

2

2e-3

90k

12 hr

54.2

79.6

59.2

NOTE. In the above tables, all models have been run at least two times with close results. The training is relatively stable. RPN means Faster R-CNN. The annotations of PASCAL VOC are transformed to COCO format, and COCO API was used for evaluation. The results are different from the official VOC evaluation. If you want to compare the VOC results in publication, please use the official VOC code for evaluation.

Requirements

NVIDIA GPU and cuDNN are required to have fast speeds. For now, CUDA 8.0 with cuDNN 6.0.20 has been tested. The other versions should be working.

Caffe MATLAB wrapper is required to run the detection/evaluation demo.

Installation

Clone the Cascade-RCNN repository, and we'll call the directory that you cloned Cascade-RCNN into CASCADE_ROOT

git clone https://github.com/zhaoweicai/cascade-rcnn.git

Build Cascade-RCNN

cd $CASCADE_ROOT/

# Follow the Caffe installation instructions here:

# http://caffe.berkeleyvision.org/installation.html

# If you're experienced with Caffe and have all of the requirements installed

# and your Makefile.config in place, then simply do:

make all -j 16

# If you want to run Cascade-RCNN detection/evaluation demo, build MATLAB wrapper as well

make matcaffe

Datasets

If you already have a COCO/VOC copy but not as organized as below, you can simply create Symlinks to have the same directory structure.

MS-COCO

In all MS-COCO experiments, we use train2017 for training, and val2017 (a.k.a. minival) for validation. Follow MS-COCO website to download images/annotations, and set-up the COCO API.

Assumed that your local COCO dataset copy is at /your/path/to/coco, make sure it has the following directory structure:

coco

|_ images

|_ train2017

| |_ .jpg

| |_ ...

| |_ .jpg

|_ val2017

|_ ...

|_ annotations

|_ instances_train2017.json

|_ instances_val2017.json

|_ ...

|_ MatlabAPI

PASCAL VOC

In all PASCAL VOC experiments, we use VOC2007+VOC2012 trainval for training, and VOC2007 test for validation. Follow PASCAL VOC website to download images/annotations, and set-up the VOCdevkit.

Assumed that your local VOCdevkit copy is at /your/path/to/VOCdevkit, make sure it has the following directory structure:

VOCdevkit

|_ VOC2007

|_ JPEGImages

| |_ <000001>.jpg

| |_ ...

| |_ <009963>.jpg

|_ Annotations

| |_ <000001>.xml

| |_ ...

| |_ <009963>.xml

|_ ...

|_ VOC2012

|_ JPEGImages

| |_ <2007_000027>.jpg

| |_ ...

| |_ <2012_004331>.jpg

|_ Annotations

| |_ <2007_000027>.xml

| |_ ...

| |_ <2012_004331>.xml

|_ ...

|_ VOCcode

Training Cascade-RCNN

Get the training data

cd $CASCADE_ROOT/data/

sh get_coco_data.sh

This will download the window files required for the experiments. You can also use the provided MATLAB scripts coco_window_file.m under $CASCADE_ROOT/data/coco/ to generate your own window files.

Download the pretrained models on ImageNet. For AlexNet and VGG-Net, the FC layers are pruned and 2048 units per FC layer are remained. In addition, the two FC layers are copied three times for Cascade R-CNN training. For ResNet, the BatchNorm layers are merged into Scale layers and frozen during training as common practice.

cd $CASCADE_ROOT/models/

sh fetch_vggnet.sh

Multiple shell scripts are provided to train Cascade-RCNN on different baseline detectors as described in our paper. Under each model folder, you need to change the root_folder of the data layer in train.prototxt and test.prototxt to your COCO path. After that, you can start to train your own Cascade-RCNN models. Take vgg-12s-600-rpn-cascade for example.

cd $CASCADE_ROOT/examples/coco/vgg-12s-600-rpn-cascade/

sh train_detection.sh

Log file will be generated along the training procedure. The total training time depends on the complexity of models and datasets. If you want to quickly check if the training works well, try the light AlexNet model on VOC dataset.

NOTE. Occasionally, the training of the Res101-FPN-Cascade will be out of memory. Just resume the training from the latest solverstate.

Pretrained Models

We only provide the Res50-FPN-baseline, Res50-FPN-Cascade and Res101-FPN-Cascade models for COCO dataset, and Res101-RFCN-Cascade for VOC dataset.

Download pre-trained models

cd $CASCADE_ROOT/examples/coco/

sh fetch_cascadercnn_models.sh

The pretrained models produce exactly the same results as described in our paper.

Testing/Evaluation Demo

Once the models pretrained or trained by yourself are available, you can use the MATLAB script run_cascadercnn_coco.m to obtain the detection and evaluation results. Set the right dataset path and choose the model of your interest to test in the demo script. The default setting is for the pretrained model. The final detection results will be saved under $CASCADE_ROOT/examples/coco/detections/ and the evaluation results will be saved under the model folder.

You also can run the shell script test_coco_detection.sh under each model folder for evalution, but it is not identical to the official evaluation. For publication, use the MATLAB script.

Disclaimer

When we were re-implementing the FPN framework and roi_align layer, we only referred to their published papers. Thus, our implementation details could be different from the official Detectron.

If you encounter any issue when using our code or model, please let me know.

 类似资料: